var/home/core/zuul-output/0000755000175000017500000000000015115561731014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015115572704015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004630332415115572676017721 0ustar rootrootDec 08 14:46:17 crc systemd[1]: Starting Kubernetes Kubelet... Dec 08 14:46:17 crc restorecon[4697]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 08 14:46:17 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 08 14:46:18 crc restorecon[4697]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 08 14:46:19 crc kubenswrapper[4894]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 14:46:19 crc kubenswrapper[4894]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 08 14:46:19 crc kubenswrapper[4894]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 14:46:19 crc kubenswrapper[4894]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 14:46:19 crc kubenswrapper[4894]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 08 14:46:19 crc kubenswrapper[4894]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.023830 4894 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026613 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026632 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026636 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026641 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026644 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026648 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026653 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026657 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026661 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026666 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026671 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026685 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026690 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026694 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026697 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026701 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026706 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026710 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026714 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026718 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026721 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026725 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026728 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026733 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026738 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026742 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026747 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026751 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026755 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026758 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026762 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026766 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026769 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026773 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026777 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026781 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026785 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026789 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026793 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026797 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026802 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026806 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026809 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026829 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026834 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026838 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026841 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026845 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026850 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026854 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026858 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026861 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026865 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026868 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026872 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026875 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026879 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026882 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026886 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026889 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026893 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026896 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026900 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026903 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026908 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026911 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026915 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026919 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026923 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026927 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.026932 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027347 4894 flags.go:64] FLAG: --address="0.0.0.0" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027358 4894 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027367 4894 flags.go:64] FLAG: --anonymous-auth="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027373 4894 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027379 4894 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027386 4894 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027392 4894 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027397 4894 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027401 4894 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027405 4894 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027410 4894 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027414 4894 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027418 4894 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027422 4894 flags.go:64] FLAG: --cgroup-root="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027426 4894 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027430 4894 flags.go:64] FLAG: --client-ca-file="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027434 4894 flags.go:64] FLAG: --cloud-config="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027438 4894 flags.go:64] FLAG: --cloud-provider="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027441 4894 flags.go:64] FLAG: --cluster-dns="[]" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027447 4894 flags.go:64] FLAG: --cluster-domain="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027450 4894 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027455 4894 flags.go:64] FLAG: --config-dir="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027459 4894 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027463 4894 flags.go:64] FLAG: --container-log-max-files="5" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027469 4894 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027473 4894 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027477 4894 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027481 4894 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027485 4894 flags.go:64] FLAG: --contention-profiling="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027491 4894 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027495 4894 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027500 4894 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027504 4894 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027509 4894 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027513 4894 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027517 4894 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027521 4894 flags.go:64] FLAG: --enable-load-reader="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027527 4894 flags.go:64] FLAG: --enable-server="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027532 4894 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027536 4894 flags.go:64] FLAG: --event-burst="100" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027541 4894 flags.go:64] FLAG: --event-qps="50" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027545 4894 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027549 4894 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027553 4894 flags.go:64] FLAG: --eviction-hard="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027558 4894 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027562 4894 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027566 4894 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027571 4894 flags.go:64] FLAG: --eviction-soft="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027575 4894 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027579 4894 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027583 4894 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027586 4894 flags.go:64] FLAG: --experimental-mounter-path="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027590 4894 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027594 4894 flags.go:64] FLAG: --fail-swap-on="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027598 4894 flags.go:64] FLAG: --feature-gates="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027603 4894 flags.go:64] FLAG: --file-check-frequency="20s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027607 4894 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027611 4894 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027615 4894 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027619 4894 flags.go:64] FLAG: --healthz-port="10248" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027623 4894 flags.go:64] FLAG: --help="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027627 4894 flags.go:64] FLAG: --hostname-override="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027631 4894 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027635 4894 flags.go:64] FLAG: --http-check-frequency="20s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027639 4894 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027643 4894 flags.go:64] FLAG: --image-credential-provider-config="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027647 4894 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027651 4894 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027655 4894 flags.go:64] FLAG: --image-service-endpoint="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027660 4894 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027664 4894 flags.go:64] FLAG: --kube-api-burst="100" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027668 4894 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027672 4894 flags.go:64] FLAG: --kube-api-qps="50" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027677 4894 flags.go:64] FLAG: --kube-reserved="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027685 4894 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027689 4894 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027693 4894 flags.go:64] FLAG: --kubelet-cgroups="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027697 4894 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027701 4894 flags.go:64] FLAG: --lock-file="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027705 4894 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027709 4894 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027713 4894 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027719 4894 flags.go:64] FLAG: --log-json-split-stream="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027723 4894 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027727 4894 flags.go:64] FLAG: --log-text-split-stream="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027731 4894 flags.go:64] FLAG: --logging-format="text" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027735 4894 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027739 4894 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027743 4894 flags.go:64] FLAG: --manifest-url="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027747 4894 flags.go:64] FLAG: --manifest-url-header="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027752 4894 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027756 4894 flags.go:64] FLAG: --max-open-files="1000000" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027761 4894 flags.go:64] FLAG: --max-pods="110" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027766 4894 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027770 4894 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027774 4894 flags.go:64] FLAG: --memory-manager-policy="None" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027778 4894 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027782 4894 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027786 4894 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027791 4894 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027801 4894 flags.go:64] FLAG: --node-status-max-images="50" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027805 4894 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027809 4894 flags.go:64] FLAG: --oom-score-adj="-999" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027829 4894 flags.go:64] FLAG: --pod-cidr="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027833 4894 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027840 4894 flags.go:64] FLAG: --pod-manifest-path="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027844 4894 flags.go:64] FLAG: --pod-max-pids="-1" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027849 4894 flags.go:64] FLAG: --pods-per-core="0" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027853 4894 flags.go:64] FLAG: --port="10250" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027857 4894 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027862 4894 flags.go:64] FLAG: --provider-id="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027865 4894 flags.go:64] FLAG: --qos-reserved="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027869 4894 flags.go:64] FLAG: --read-only-port="10255" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027874 4894 flags.go:64] FLAG: --register-node="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027878 4894 flags.go:64] FLAG: --register-schedulable="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027882 4894 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027889 4894 flags.go:64] FLAG: --registry-burst="10" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027893 4894 flags.go:64] FLAG: --registry-qps="5" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027896 4894 flags.go:64] FLAG: --reserved-cpus="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027900 4894 flags.go:64] FLAG: --reserved-memory="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027905 4894 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027909 4894 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027914 4894 flags.go:64] FLAG: --rotate-certificates="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027919 4894 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027924 4894 flags.go:64] FLAG: --runonce="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027930 4894 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027934 4894 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027938 4894 flags.go:64] FLAG: --seccomp-default="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027942 4894 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027946 4894 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027951 4894 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027955 4894 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027959 4894 flags.go:64] FLAG: --storage-driver-password="root" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027963 4894 flags.go:64] FLAG: --storage-driver-secure="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027967 4894 flags.go:64] FLAG: --storage-driver-table="stats" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027971 4894 flags.go:64] FLAG: --storage-driver-user="root" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027975 4894 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027979 4894 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027983 4894 flags.go:64] FLAG: --system-cgroups="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027987 4894 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027993 4894 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.027997 4894 flags.go:64] FLAG: --tls-cert-file="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028001 4894 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028006 4894 flags.go:64] FLAG: --tls-min-version="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028010 4894 flags.go:64] FLAG: --tls-private-key-file="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028014 4894 flags.go:64] FLAG: --topology-manager-policy="none" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028019 4894 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028023 4894 flags.go:64] FLAG: --topology-manager-scope="container" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028027 4894 flags.go:64] FLAG: --v="2" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028032 4894 flags.go:64] FLAG: --version="false" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028038 4894 flags.go:64] FLAG: --vmodule="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028043 4894 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028047 4894 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028162 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028167 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028171 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028175 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028179 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028183 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028187 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028190 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028195 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028199 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028202 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028207 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028211 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028214 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028218 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028221 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028225 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028228 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028232 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028235 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028239 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028242 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028246 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028249 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028253 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028256 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028260 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028263 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028266 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028271 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028275 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028278 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028282 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028286 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028290 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028293 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028302 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028307 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028312 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028317 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028322 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028327 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028340 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028344 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028349 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028353 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028357 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028361 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028365 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028368 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028372 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028375 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028379 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028382 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028385 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028389 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028392 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028396 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028399 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028402 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028406 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028409 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028414 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028418 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028422 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028426 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028431 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028434 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028440 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028444 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.028449 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.028456 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.042293 4894 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.042369 4894 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042667 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042691 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042704 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042716 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042727 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042738 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042749 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042762 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042888 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042924 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042936 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.042947 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043033 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043045 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043057 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043069 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043079 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043091 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043115 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043129 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043141 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043166 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043531 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043550 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043562 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043572 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043587 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043600 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043612 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043622 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043633 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043643 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043653 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043663 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043673 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043683 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043693 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043706 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043718 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043729 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043742 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043753 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043764 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043775 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043786 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043800 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043811 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043852 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043864 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043874 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043885 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043910 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043923 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043933 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043943 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043952 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043960 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043968 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043976 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043984 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043991 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.043999 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044007 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044016 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044024 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044032 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044040 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044047 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044055 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044064 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044072 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.044086 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044330 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044345 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044354 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044365 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044380 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044390 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044399 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044409 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044418 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044426 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044434 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044443 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044451 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044460 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044468 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044477 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044487 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044496 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044506 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044514 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044523 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044532 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044540 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044549 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044557 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044566 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044574 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044584 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044595 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044603 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044611 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044619 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044627 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044635 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044643 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044650 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044659 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044667 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044675 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044684 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044691 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044699 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044707 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044714 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044723 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044733 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044743 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044754 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044764 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044775 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044786 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044796 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044805 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044880 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044892 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044903 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044913 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044923 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044933 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044943 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044953 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044962 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044970 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044978 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044985 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.044994 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.045001 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.045010 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.045017 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.045025 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.045034 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.045046 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.045344 4894 server.go:940] "Client rotation is on, will bootstrap in background" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.050091 4894 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.050242 4894 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.051189 4894 server.go:997] "Starting client certificate rotation" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.051232 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.051442 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-03 21:06:15.657118475 +0000 UTC Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.051568 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.059875 4894 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.062401 4894 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.063572 4894 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.078936 4894 log.go:25] "Validated CRI v1 runtime API" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.107792 4894 log.go:25] "Validated CRI v1 image API" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.110517 4894 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.113522 4894 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-08-14-41-39-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.113557 4894 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.132718 4894 manager.go:217] Machine: {Timestamp:2025-12-08 14:46:19.131263593 +0000 UTC m=+0.231269718 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0f545b9e-425e-49eb-9611-92fe15e98ca0 BootID:058a8d66-2b8f-482d-b95a-6b5107a7954b Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7a:27:2f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:7a:27:2f Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:04:d9:2b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:bb:96:35 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c4:2b:29 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:15:36:f7 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4a:ef:d9:21:65:ac Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:de:b0:5d:ff:ba:dc Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.133286 4894 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.133596 4894 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.134291 4894 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.134545 4894 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.134630 4894 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.134926 4894 topology_manager.go:138] "Creating topology manager with none policy" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.135017 4894 container_manager_linux.go:303] "Creating device plugin manager" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.135255 4894 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.135347 4894 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.135726 4894 state_mem.go:36] "Initialized new in-memory state store" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.135881 4894 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.136785 4894 kubelet.go:418] "Attempting to sync node with API server" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.136913 4894 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.136994 4894 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.137063 4894 kubelet.go:324] "Adding apiserver pod source" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.137119 4894 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.138763 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.138812 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.138933 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.138878 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.139190 4894 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.139548 4894 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.140572 4894 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141124 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141146 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141154 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141161 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141173 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141180 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141194 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141205 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141216 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141224 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141252 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141259 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141445 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.141862 4894 server.go:1280] "Started kubelet" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.142221 4894 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.142228 4894 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.142642 4894 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 08 14:46:19 crc systemd[1]: Started Kubernetes Kubelet. Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.143883 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.143750 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.51:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f44c20a6dd2c1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-08 14:46:19.141837505 +0000 UTC m=+0.241843630,LastTimestamp:2025-12-08 14:46:19.141837505 +0000 UTC m=+0.241843630,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.145420 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.145579 4894 server.go:460] "Adding debug handlers to kubelet server" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.145581 4894 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.146331 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 15:31:42.919539327 +0000 UTC Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.146412 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 864h45m23.773129697s for next certificate rotation Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.146531 4894 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.148113 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="200ms" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.147134 4894 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.148246 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.148309 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.146541 4894 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.148981 4894 factory.go:153] Registering CRI-O factory Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.148998 4894 factory.go:221] Registration of the crio container factory successfully Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.149059 4894 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.149067 4894 factory.go:55] Registering systemd factory Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.149072 4894 factory.go:221] Registration of the systemd container factory successfully Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.149090 4894 factory.go:103] Registering Raw factory Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.149101 4894 manager.go:1196] Started watching for new ooms in manager Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.149585 4894 manager.go:319] Starting recovery of all containers Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.148981 4894 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165527 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165620 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165632 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165643 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165653 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165663 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165673 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165682 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165695 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165705 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165714 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165722 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165734 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165744 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165752 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165760 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165769 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165777 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165787 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165796 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165805 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165853 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165862 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165886 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165909 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165918 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165928 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165938 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165947 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.165959 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166026 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166037 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166045 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166055 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166064 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166089 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166099 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166119 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166127 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166137 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166146 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166155 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166165 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166175 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166184 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166195 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166204 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166214 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166223 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166232 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166241 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166252 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166265 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166276 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166286 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166296 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166306 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166314 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166324 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166332 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166342 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166352 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166360 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166368 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166376 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166383 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166391 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166400 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166409 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166417 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166425 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166433 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166443 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166451 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166461 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166472 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166481 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166490 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166501 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166512 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166521 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166533 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166543 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166551 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166560 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166569 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166578 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166587 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166599 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166610 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166619 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166627 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166636 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166646 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166655 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166663 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166672 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166682 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166690 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166699 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166709 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166717 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166726 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166734 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166748 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166758 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166768 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166779 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166788 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166798 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166808 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166832 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166842 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166851 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166858 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166868 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166876 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166886 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166895 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166904 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166915 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166925 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166933 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166942 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166951 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166962 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166971 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166980 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166989 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.166998 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167006 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167015 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167024 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167034 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167043 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167054 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167065 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167074 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167083 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167092 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167101 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167112 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167848 4894 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167867 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167877 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167915 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167926 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167937 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167947 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167956 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167966 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167976 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167986 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.167996 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168006 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168015 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168025 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168036 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168046 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168055 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168065 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168074 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168085 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168094 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168104 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168112 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168122 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168133 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168144 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168154 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168163 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168173 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168183 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168194 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168204 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168215 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168225 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168235 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168245 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168255 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168266 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168276 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168286 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168295 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168305 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168316 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168325 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168335 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168345 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168357 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168366 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168374 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168384 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168393 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168403 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168414 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168424 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168434 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168444 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168455 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168465 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168475 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168484 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168494 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168503 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168517 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168533 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168542 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168550 4894 reconstruct.go:97] "Volume reconstruction finished" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.168558 4894 reconciler.go:26] "Reconciler: start to sync state" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.176973 4894 manager.go:324] Recovery completed Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.189738 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.191370 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.191429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.191438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.192568 4894 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.192582 4894 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.192602 4894 state_mem.go:36] "Initialized new in-memory state store" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.192908 4894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.195435 4894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.195496 4894 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.195542 4894 kubelet.go:2335] "Starting kubelet main sync loop" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.195783 4894 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.197046 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.197132 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.201923 4894 policy_none.go:49] "None policy: Start" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.203215 4894 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.203238 4894 state_mem.go:35] "Initializing new in-memory state store" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.248420 4894 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.257322 4894 manager.go:334] "Starting Device Plugin manager" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.257396 4894 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.257411 4894 server.go:79] "Starting device plugin registration server" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.257898 4894 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.257915 4894 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.258461 4894 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.258578 4894 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.258599 4894 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.268679 4894 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.296886 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.296986 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.298307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.298365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.298376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.298559 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.298748 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.298796 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302181 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302369 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302839 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302951 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.302995 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305477 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305745 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305901 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.305932 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.306783 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.306807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.306828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.306923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.306942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.306945 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.306952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.307141 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.307215 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.307465 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.307492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.307502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.307689 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.307719 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.308472 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.308519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.308539 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.308865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.308887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.308896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.348681 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="400ms" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.361153 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.364043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.364463 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.364619 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.364743 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.365484 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.370366 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.370567 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.370679 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.370826 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.370937 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371037 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371139 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371366 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371448 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371613 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371723 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371856 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.371972 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.372074 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.472941 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473069 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473101 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473161 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473190 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473338 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473258 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473179 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473451 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473471 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473225 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473402 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473503 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473557 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473582 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473618 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473623 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473654 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473659 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473687 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473688 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473720 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.473766 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.565664 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.567471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.567529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.567548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.567581 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.568265 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.637321 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.656493 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.670765 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.678045 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-da97a86577934fb47b02ba96c07c55d0498173a09ad540010dc3d3f02dcec9e0 WatchSource:0}: Error finding container da97a86577934fb47b02ba96c07c55d0498173a09ad540010dc3d3f02dcec9e0: Status 404 returned error can't find the container with id da97a86577934fb47b02ba96c07c55d0498173a09ad540010dc3d3f02dcec9e0 Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.680501 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-44de9f9072864754297d9989d1ae376754ba366ef817ab27307101e3e9010e04 WatchSource:0}: Error finding container 44de9f9072864754297d9989d1ae376754ba366ef817ab27307101e3e9010e04: Status 404 returned error can't find the container with id 44de9f9072864754297d9989d1ae376754ba366ef817ab27307101e3e9010e04 Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.690708 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-1132aaad3331af1b3c04198d08edd3655940b5dfc06de9860b2444dcdde2d541 WatchSource:0}: Error finding container 1132aaad3331af1b3c04198d08edd3655940b5dfc06de9860b2444dcdde2d541: Status 404 returned error can't find the container with id 1132aaad3331af1b3c04198d08edd3655940b5dfc06de9860b2444dcdde2d541 Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.692501 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.696861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.723371 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0f0c1acf44222ccbf1bee14dd7b7194a9b9951a07dc27f75ac7aaaf40e8c9da0 WatchSource:0}: Error finding container 0f0c1acf44222ccbf1bee14dd7b7194a9b9951a07dc27f75ac7aaaf40e8c9da0: Status 404 returned error can't find the container with id 0f0c1acf44222ccbf1bee14dd7b7194a9b9951a07dc27f75ac7aaaf40e8c9da0 Dec 08 14:46:19 crc kubenswrapper[4894]: W1208 14:46:19.731460 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f653ac37bceb5335813be061d70ae5798ac10bcf49f899534b55022f15a286b4 WatchSource:0}: Error finding container f653ac37bceb5335813be061d70ae5798ac10bcf49f899534b55022f15a286b4: Status 404 returned error can't find the container with id f653ac37bceb5335813be061d70ae5798ac10bcf49f899534b55022f15a286b4 Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.749392 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="800ms" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.968891 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.970854 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.970899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.970912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:19 crc kubenswrapper[4894]: I1208 14:46:19.970942 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 14:46:19 crc kubenswrapper[4894]: E1208 14:46:19.971491 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Dec 08 14:46:20 crc kubenswrapper[4894]: W1208 14:46:20.124213 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:20 crc kubenswrapper[4894]: E1208 14:46:20.124313 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.144758 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.200463 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3" exitCode=0 Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.200536 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.200654 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f653ac37bceb5335813be061d70ae5798ac10bcf49f899534b55022f15a286b4"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.200841 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.201893 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.201942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.201955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.202103 4894 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2" exitCode=0 Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.202152 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.202167 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0f0c1acf44222ccbf1bee14dd7b7194a9b9951a07dc27f75ac7aaaf40e8c9da0"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.202218 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.202725 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.202750 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.202759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.203729 4894 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606" exitCode=0 Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.203800 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.203861 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1132aaad3331af1b3c04198d08edd3655940b5dfc06de9860b2444dcdde2d541"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.203976 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.205464 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.205493 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"da97a86577934fb47b02ba96c07c55d0498173a09ad540010dc3d3f02dcec9e0"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.206380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.206412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.206467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.207718 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971" exitCode=0 Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.207773 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.207800 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"44de9f9072864754297d9989d1ae376754ba366ef817ab27307101e3e9010e04"} Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.207939 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.208978 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.209010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.209024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.210284 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.210910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.210942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.210953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:20 crc kubenswrapper[4894]: W1208 14:46:20.533046 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:20 crc kubenswrapper[4894]: E1208 14:46:20.533491 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:20 crc kubenswrapper[4894]: E1208 14:46:20.551027 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="1.6s" Dec 08 14:46:20 crc kubenswrapper[4894]: W1208 14:46:20.657219 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:20 crc kubenswrapper[4894]: E1208 14:46:20.657332 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:20 crc kubenswrapper[4894]: W1208 14:46:20.702160 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:20 crc kubenswrapper[4894]: E1208 14:46:20.702280 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.772121 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.773720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.773776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.773788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:20 crc kubenswrapper[4894]: I1208 14:46:20.773832 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 14:46:20 crc kubenswrapper[4894]: E1208 14:46:20.775066 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.51:6443: connect: connection refused" node="crc" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.065423 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 08 14:46:21 crc kubenswrapper[4894]: E1208 14:46:21.066613 4894 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.51:6443: connect: connection refused" logger="UnhandledError" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.145459 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.51:6443: connect: connection refused Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.211897 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"48ebdc46c969cd502cf4084c56a87b1129f69d0c8b3da96efc447f6be04be840"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.212006 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.213673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.213729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.213741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.216339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.216364 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.216376 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.216440 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.217127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.217156 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.217167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.219771 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.219800 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.219844 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.219890 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.221179 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.221204 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.221212 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.224511 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.224533 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.224545 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.224555 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.226438 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a" exitCode=0 Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.226458 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a"} Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.226629 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.227319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.227345 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.227353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.248952 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:21 crc kubenswrapper[4894]: I1208 14:46:21.871788 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.231935 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43" exitCode=0 Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.232010 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43"} Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.232155 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.233027 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.233056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.233103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.236968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8"} Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.237061 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.237378 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.238104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.238130 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.238164 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.239851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.239896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.239914 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.321314 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.321549 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.323189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.323249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.323263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.375327 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.377058 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.377144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.377159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.377217 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 14:46:22 crc kubenswrapper[4894]: I1208 14:46:22.615264 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.246301 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc"} Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.246389 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1"} Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.246415 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75"} Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.246433 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448"} Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.246469 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.246489 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.246568 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.248039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.248077 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.248087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.249104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.249136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.249145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.965634 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:23 crc kubenswrapper[4894]: I1208 14:46:23.974641 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.249118 4894 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.249183 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.254665 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b"} Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.254757 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.254807 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.254954 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.255897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.255918 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.255923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.255988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.255945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.256023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.256371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.256412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:24 crc kubenswrapper[4894]: I1208 14:46:24.256428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.144285 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.257094 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.257149 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.257169 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.258669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.258705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.258715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.258760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.258793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.258848 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.259655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.259721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.259741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:25 crc kubenswrapper[4894]: I1208 14:46:25.325728 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 08 14:46:26 crc kubenswrapper[4894]: I1208 14:46:26.586580 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 08 14:46:26 crc kubenswrapper[4894]: I1208 14:46:26.586873 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:26 crc kubenswrapper[4894]: I1208 14:46:26.588373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:26 crc kubenswrapper[4894]: I1208 14:46:26.588444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:26 crc kubenswrapper[4894]: I1208 14:46:26.588468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:27 crc kubenswrapper[4894]: I1208 14:46:27.907776 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:27 crc kubenswrapper[4894]: I1208 14:46:27.908016 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:27 crc kubenswrapper[4894]: I1208 14:46:27.912172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:27 crc kubenswrapper[4894]: I1208 14:46:27.912225 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:27 crc kubenswrapper[4894]: I1208 14:46:27.912235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:29 crc kubenswrapper[4894]: I1208 14:46:29.230386 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 08 14:46:29 crc kubenswrapper[4894]: I1208 14:46:29.230646 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:29 crc kubenswrapper[4894]: I1208 14:46:29.232036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:29 crc kubenswrapper[4894]: I1208 14:46:29.232114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:29 crc kubenswrapper[4894]: I1208 14:46:29.232134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:29 crc kubenswrapper[4894]: E1208 14:46:29.268870 4894 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 08 14:46:31 crc kubenswrapper[4894]: I1208 14:46:31.033143 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 08 14:46:31 crc kubenswrapper[4894]: I1208 14:46:31.033274 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 08 14:46:31 crc kubenswrapper[4894]: I1208 14:46:31.876673 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:31 crc kubenswrapper[4894]: I1208 14:46:31.876835 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:31 crc kubenswrapper[4894]: I1208 14:46:31.878146 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:31 crc kubenswrapper[4894]: I1208 14:46:31.878204 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:31 crc kubenswrapper[4894]: I1208 14:46:31.878218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:32 crc kubenswrapper[4894]: I1208 14:46:32.084445 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 08 14:46:32 crc kubenswrapper[4894]: I1208 14:46:32.084531 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 08 14:46:32 crc kubenswrapper[4894]: I1208 14:46:32.092399 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 08 14:46:32 crc kubenswrapper[4894]: I1208 14:46:32.092492 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 08 14:46:32 crc kubenswrapper[4894]: I1208 14:46:32.625887 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]log ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]etcd ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/generic-apiserver-start-informers ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-filter ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-apiextensions-informers ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-apiextensions-controllers ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/crd-informer-synced ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-system-namespaces-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 08 14:46:32 crc kubenswrapper[4894]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 08 14:46:32 crc kubenswrapper[4894]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/bootstrap-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/start-kube-aggregator-informers ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/apiservice-registration-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/apiservice-discovery-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]autoregister-completion ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/apiservice-openapi-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 08 14:46:32 crc kubenswrapper[4894]: livez check failed Dec 08 14:46:32 crc kubenswrapper[4894]: I1208 14:46:32.626483 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:46:34 crc kubenswrapper[4894]: I1208 14:46:34.249167 4894 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 08 14:46:34 crc kubenswrapper[4894]: I1208 14:46:34.249247 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.085697 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.093006 4894 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.093089 4894 trace.go:236] Trace[1039298838]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 14:46:23.850) (total time: 13242ms): Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[1039298838]: ---"Objects listed" error: 13242ms (14:46:37.093) Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[1039298838]: [13.242824499s] [13.242824499s] END Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.093344 4894 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.093123 4894 trace.go:236] Trace[2066123344]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 14:46:23.125) (total time: 13967ms): Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[2066123344]: ---"Objects listed" error: 13967ms (14:46:37.093) Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[2066123344]: [13.967495251s] [13.967495251s] END Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.093437 4894 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.094333 4894 trace.go:236] Trace[1230461194]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 14:46:23.304) (total time: 13789ms): Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[1230461194]: ---"Objects listed" error: 13789ms (14:46:37.094) Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[1230461194]: [13.789888076s] [13.789888076s] END Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.094364 4894 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.095445 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.097515 4894 trace.go:236] Trace[1984205242]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Dec-2025 14:46:22.527) (total time: 14569ms): Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[1984205242]: ---"Objects listed" error: 14569ms (14:46:37.097) Dec 08 14:46:37 crc kubenswrapper[4894]: Trace[1984205242]: [14.569884902s] [14.569884902s] END Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.097534 4894 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.099356 4894 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.125684 4894 csr.go:261] certificate signing request csr-klnjc is approved, waiting to be issued Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.131991 4894 csr.go:257] certificate signing request csr-klnjc is issued Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.144575 4894 apiserver.go:52] "Watching apiserver" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.150393 4894 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.150649 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.151078 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.151204 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.151088 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.151293 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.151451 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.151825 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.151866 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.152037 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.152086 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.153688 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.153893 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.155945 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.155958 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.156981 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.157113 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.157519 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.157773 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.157959 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.193167 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.208440 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.220951 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.231977 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.235776 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59534->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.235840 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59534->192.168.126.11:17697: read: connection reset by peer" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.245945 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.249541 4894 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.260483 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.271436 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.281679 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294104 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294145 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294185 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294203 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294221 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294257 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294278 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294332 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294350 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294383 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294403 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294433 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294467 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294484 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294502 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294519 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294540 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294558 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294577 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294629 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294646 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294663 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294675 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294729 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294714 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294774 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294838 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294861 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294854 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294853 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294903 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294880 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.294990 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295015 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295035 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295068 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295079 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295192 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295199 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295277 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295258 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295299 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295323 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295340 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295360 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295380 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295401 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295420 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295442 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295460 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295476 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295493 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295510 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295529 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295546 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295555 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295562 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295595 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295586 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295638 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295730 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295686 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295769 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295807 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295860 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295868 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295892 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295884 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295952 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.295986 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296094 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296115 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296126 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296141 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296149 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296156 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296167 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296203 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296232 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296255 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296281 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296305 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296329 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296440 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296466 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296490 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296515 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296565 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296591 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296620 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296653 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296682 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296700 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296710 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296739 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296769 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296801 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296861 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296899 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296936 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296966 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296998 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297035 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297063 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297097 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297130 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297163 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297197 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297227 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297257 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297291 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297439 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297510 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297547 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297581 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297617 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297649 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297683 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297712 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297740 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297771 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297851 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297886 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297918 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297950 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297982 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298013 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298046 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298079 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298134 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298165 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298195 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298226 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298261 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298291 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298323 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298358 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298389 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298418 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298448 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298480 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298512 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298543 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298578 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298615 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298644 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298720 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298761 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299071 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299104 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299140 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299180 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299215 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299290 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300561 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300620 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300655 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8" exitCode=255 Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300697 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8"} Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300802 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300852 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300878 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301010 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301035 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301060 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301084 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301108 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301142 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301315 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301384 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301420 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301454 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301484 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301526 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301559 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301592 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301626 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301659 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301727 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301761 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301878 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301912 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301944 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302097 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302153 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302187 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302222 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302257 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302329 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302366 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302402 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302437 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302472 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302505 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302539 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302572 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302649 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302687 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302726 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302764 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302801 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302880 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302935 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302962 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.302985 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303012 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303037 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303072 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303116 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303182 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303267 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303304 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303339 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303415 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303454 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303504 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303538 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303573 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303608 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.303786 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.305949 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296273 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.321872 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296336 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296798 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.296984 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297135 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297318 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297403 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297470 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297517 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297767 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297970 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.297999 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298021 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298150 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298248 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298347 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298367 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298562 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298737 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.298775 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299132 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299648 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299668 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299851 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.299965 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300341 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300455 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300495 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300543 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.300924 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.301065 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.306444 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.306919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.307001 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.307223 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.307244 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.307446 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.308736 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.308770 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.309724 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.310145 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.310436 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.310586 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.310912 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.310925 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.311139 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.311615 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.312481 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.313009 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.313903 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.314098 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.314257 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322732 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.314899 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.315632 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.315699 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.316176 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.316411 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.316625 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.316872 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.317110 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.317882 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.317969 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.318561 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.318569 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.318558 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.318891 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319019 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319175 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319314 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319466 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319503 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319746 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319930 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319957 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.320001 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.320174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.320426 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.320442 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.320681 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.320857 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.321390 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.321439 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.321577 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.321781 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.321904 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322024 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323124 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322355 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.319794 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323152 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323197 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322429 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322485 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322508 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323255 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323083 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.322304 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323119 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323241 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323320 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323292 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323601 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323723 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.324088 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.325709 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.326347 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.326382 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.326756 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.326844 4894 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327071 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.327188 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327269 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327383 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327471 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327606 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.327624 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327876 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327941 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.327997 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328021 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328112 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328253 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328266 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328298 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328310 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328337 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328470 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328624 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328634 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.328912 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.329793 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.329801 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330053 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330130 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330326 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330400 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330784 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330859 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330894 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.330935 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.331328 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.331472 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.331486 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.331799 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.331849 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.331875 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.332195 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.332287 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.332459 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.332867 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.332928 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.333204 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.333790 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334209 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334276 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334254 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334361 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.334403 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:46:37.823405907 +0000 UTC m=+18.923412022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334485 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334625 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334773 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334849 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334880 4894 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334901 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334930 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334948 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334965 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334981 4894 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.334997 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335015 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335032 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335048 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335061 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335076 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335091 4894 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335106 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.335163 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.335636 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:37.83560923 +0000 UTC m=+18.935615345 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.335684 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:37.835675051 +0000 UTC m=+18.935681156 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.336433 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.336683 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.323965 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.341445 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.343141 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.342958 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.343422 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.343599 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.344472 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.345067 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.349785 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.351759 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.351893 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.352064 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.355289 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.357174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.358475 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.358493 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.358631 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.359452 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.359485 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.359504 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.359601 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:37.859572061 +0000 UTC m=+18.959578386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.360926 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.361036 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.361067 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.361088 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.361157 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:37.861130736 +0000 UTC m=+18.961136871 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.362490 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.362577 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.363284 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.363459 4894 scope.go:117] "RemoveContainer" containerID="8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.366850 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.375154 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.375685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.387679 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.388230 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.389441 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.391576 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.405160 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.415434 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440270 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440319 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440332 4894 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440343 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440352 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440361 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440369 4894 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440378 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440389 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440411 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440423 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440434 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440443 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440453 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440462 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440470 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440480 4894 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440490 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440499 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440508 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440517 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440526 4894 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440590 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440374 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.440739 4894 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441153 4894 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441259 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441271 4894 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441282 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441318 4894 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441364 4894 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441373 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441384 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441394 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441633 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441647 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441656 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441665 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441711 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441721 4894 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441730 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441739 4894 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441747 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441756 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441765 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441907 4894 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441918 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441926 4894 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441956 4894 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441965 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441974 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441983 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.441991 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442000 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442009 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442018 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442026 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442581 4894 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442598 4894 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442607 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442617 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442625 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442634 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442642 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442650 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442658 4894 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442667 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442675 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442684 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442692 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442699 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442708 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442715 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442723 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442732 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442741 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442748 4894 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442756 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442765 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442774 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442783 4894 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442792 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442801 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442810 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442844 4894 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442853 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442861 4894 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442870 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442879 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442887 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442895 4894 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442903 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442911 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442919 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442927 4894 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442935 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442943 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442950 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442958 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442966 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442973 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442981 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442988 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.442996 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443004 4894 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443012 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443021 4894 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443030 4894 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443038 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443049 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443057 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443065 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443073 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443083 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443090 4894 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443098 4894 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443107 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443114 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443122 4894 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443130 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443137 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443145 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443153 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443160 4894 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443168 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443175 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443197 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443206 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443214 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443221 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443229 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443237 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443245 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443252 4894 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443260 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443268 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443276 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443283 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443291 4894 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443300 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443309 4894 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443317 4894 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443324 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443332 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443340 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443348 4894 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443356 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443364 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443372 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443379 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443387 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443394 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443402 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443410 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443629 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443641 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443654 4894 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443663 4894 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443671 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443679 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443688 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443695 4894 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443725 4894 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443734 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443742 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443750 4894 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443758 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443766 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443776 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443785 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443793 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443802 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443829 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443838 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443847 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443855 4894 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443864 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443872 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.443881 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.473136 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.474149 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.479788 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 08 14:46:37 crc kubenswrapper[4894]: W1208 14:46:37.493494 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-382080a8295aa29508cb9f8902acdcedecb62a386d9022400c5f37b77724c81c WatchSource:0}: Error finding container 382080a8295aa29508cb9f8902acdcedecb62a386d9022400c5f37b77724c81c: Status 404 returned error can't find the container with id 382080a8295aa29508cb9f8902acdcedecb62a386d9022400c5f37b77724c81c Dec 08 14:46:37 crc kubenswrapper[4894]: W1208 14:46:37.499271 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-08b8d1c35198975bdcaf4f996dd6d9d57613b98ab9ebb8af751b8d8c2d3ec944 WatchSource:0}: Error finding container 08b8d1c35198975bdcaf4f996dd6d9d57613b98ab9ebb8af751b8d8c2d3ec944: Status 404 returned error can't find the container with id 08b8d1c35198975bdcaf4f996dd6d9d57613b98ab9ebb8af751b8d8c2d3ec944 Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.620037 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.663933 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.677492 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.693991 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.710898 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.730736 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.747242 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.760623 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.848215 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.848314 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.848352 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.848431 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:46:38.848389918 +0000 UTC m=+19.948396033 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.848468 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.848504 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.848522 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:38.848507491 +0000 UTC m=+19.948513596 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.848620 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:38.848599364 +0000 UTC m=+19.948605479 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.949073 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:37 crc kubenswrapper[4894]: I1208 14:46:37.949125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949289 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949314 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949327 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949390 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:38.949371362 +0000 UTC m=+20.049377477 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949460 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949473 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949482 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:37 crc kubenswrapper[4894]: E1208 14:46:37.949510 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:38.949501886 +0000 UTC m=+20.049508001 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.133437 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-08 14:41:37 +0000 UTC, rotation deadline is 2026-09-25 16:01:09.79039317 +0000 UTC Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.133501 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6985h14m31.656894455s for next certificate rotation Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.196235 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.196368 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.305387 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.306847 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf"} Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.307947 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.309236 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2"} Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.309267 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"228af570403f9d6e2378ea6bb61ff3aec4eb8d97be27b1cc64a856338e7487fc"} Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.311259 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1"} Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.311333 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f"} Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.311347 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"08b8d1c35198975bdcaf4f996dd6d9d57613b98ab9ebb8af751b8d8c2d3ec944"} Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.312767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"382080a8295aa29508cb9f8902acdcedecb62a386d9022400c5f37b77724c81c"} Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.314168 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.336749 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.351053 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.367390 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.371952 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-97dqr"] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.372422 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-rr2kq"] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.372596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.372663 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.374180 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.374622 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.375106 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.375326 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.375877 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.376019 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.376856 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.379317 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.386224 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.402943 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.417450 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.429677 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.443013 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.462134 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.489600 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.513755 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.526707 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.543340 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.555879 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b27019e5-2a3d-414e-b2ee-7606492ba074-rootfs\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.555942 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b27019e5-2a3d-414e-b2ee-7606492ba074-mcd-auth-proxy-config\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.555961 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6f4dd416-8f0d-4efb-944d-1720665a8d2c-hosts-file\") pod \"node-resolver-rr2kq\" (UID: \"6f4dd416-8f0d-4efb-944d-1720665a8d2c\") " pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.556005 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b27019e5-2a3d-414e-b2ee-7606492ba074-proxy-tls\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.556075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqtll\" (UniqueName: \"kubernetes.io/projected/b27019e5-2a3d-414e-b2ee-7606492ba074-kube-api-access-jqtll\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.556147 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75zfw\" (UniqueName: \"kubernetes.io/projected/6f4dd416-8f0d-4efb-944d-1720665a8d2c-kube-api-access-75zfw\") pod \"node-resolver-rr2kq\" (UID: \"6f4dd416-8f0d-4efb-944d-1720665a8d2c\") " pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.568247 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.594021 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.625991 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.657609 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b27019e5-2a3d-414e-b2ee-7606492ba074-rootfs\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.657691 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b27019e5-2a3d-414e-b2ee-7606492ba074-mcd-auth-proxy-config\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.657714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6f4dd416-8f0d-4efb-944d-1720665a8d2c-hosts-file\") pod \"node-resolver-rr2kq\" (UID: \"6f4dd416-8f0d-4efb-944d-1720665a8d2c\") " pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.657736 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b27019e5-2a3d-414e-b2ee-7606492ba074-proxy-tls\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.657757 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqtll\" (UniqueName: \"kubernetes.io/projected/b27019e5-2a3d-414e-b2ee-7606492ba074-kube-api-access-jqtll\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.657782 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75zfw\" (UniqueName: \"kubernetes.io/projected/6f4dd416-8f0d-4efb-944d-1720665a8d2c-kube-api-access-75zfw\") pod \"node-resolver-rr2kq\" (UID: \"6f4dd416-8f0d-4efb-944d-1720665a8d2c\") " pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.657788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b27019e5-2a3d-414e-b2ee-7606492ba074-rootfs\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.658179 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6f4dd416-8f0d-4efb-944d-1720665a8d2c-hosts-file\") pod \"node-resolver-rr2kq\" (UID: \"6f4dd416-8f0d-4efb-944d-1720665a8d2c\") " pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.658764 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b27019e5-2a3d-414e-b2ee-7606492ba074-mcd-auth-proxy-config\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.663221 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b27019e5-2a3d-414e-b2ee-7606492ba074-proxy-tls\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.674848 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75zfw\" (UniqueName: \"kubernetes.io/projected/6f4dd416-8f0d-4efb-944d-1720665a8d2c-kube-api-access-75zfw\") pod \"node-resolver-rr2kq\" (UID: \"6f4dd416-8f0d-4efb-944d-1720665a8d2c\") " pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.685318 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqtll\" (UniqueName: \"kubernetes.io/projected/b27019e5-2a3d-414e-b2ee-7606492ba074-kube-api-access-jqtll\") pod \"machine-config-daemon-97dqr\" (UID: \"b27019e5-2a3d-414e-b2ee-7606492ba074\") " pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.687063 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.692452 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rr2kq" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.780585 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-jtx8z"] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.780950 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-pbxln"] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.781167 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.781838 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.784390 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.786107 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.786651 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.786850 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.787027 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.787146 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.787265 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.801399 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.838181 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.860160 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.860236 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.860275 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.860341 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.860383 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:40.860369802 +0000 UTC m=+21.960375917 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.860656 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:46:40.86064763 +0000 UTC m=+21.960653745 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.860722 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.860753 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:40.860746993 +0000 UTC m=+21.960753108 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.861479 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.876173 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.894707 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.911283 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.928069 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.939718 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.954582 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.960947 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961001 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-cni-multus\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961018 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-etc-kubernetes\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961040 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68w6k\" (UniqueName: \"kubernetes.io/projected/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-kube-api-access-68w6k\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961059 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-hostroot\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-cni-bin\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961093 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-multus-certs\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961148 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-system-cni-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961156 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961197 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961309 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961353 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961361 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-conf-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-socket-dir-parent\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmnss\" (UniqueName: \"kubernetes.io/projected/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-kube-api-access-tmnss\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961479 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:40.961431798 +0000 UTC m=+22.061438063 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961510 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961533 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961535 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-os-release\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961603 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-k8s-cni-cncf-io\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961549 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961636 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cnibin\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: E1208 14:46:38.961680 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:40.961658345 +0000 UTC m=+22.061664460 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961703 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-cni-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961728 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-kubelet\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961753 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cni-binary-copy\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961777 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-cnibin\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-daemon-config\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961808 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-cni-binary-copy\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961875 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-os-release\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961895 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-system-cni-dir\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.961950 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-netns\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:38 crc kubenswrapper[4894]: I1208 14:46:38.983334 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.001496 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.023834 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.052250 4894 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052583 4894 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052640 4894 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052708 4894 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052765 4894 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052889 4894 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052892 4894 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052902 4894 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052936 4894 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052933 4894 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052977 4894 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052664 4894 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052966 4894 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.052705 4894 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.053091 4894 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.052961 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc/status\": read tcp 38.102.83.51:56662->38.102.83.51:6443: use of closed network connection" Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.053306 4894 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063139 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-cnibin\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063183 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-daemon-config\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063209 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-os-release\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063228 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-cni-binary-copy\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063247 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-system-cni-dir\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063265 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-netns\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063300 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-cni-multus\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063317 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-etc-kubernetes\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063334 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68w6k\" (UniqueName: \"kubernetes.io/projected/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-kube-api-access-68w6k\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063354 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-hostroot\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063362 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-system-cni-dir\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063429 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-netns\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063434 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-cni-bin\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063473 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-etc-kubernetes\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063477 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-cni-multus\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063372 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-cni-bin\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063534 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-multus-certs\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-system-cni-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063603 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063634 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-conf-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063656 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063678 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-socket-dir-parent\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063703 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmnss\" (UniqueName: \"kubernetes.io/projected/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-kube-api-access-tmnss\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063726 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-os-release\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063752 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-cni-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063795 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-hostroot\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063804 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-k8s-cni-cncf-io\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063863 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cnibin\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063891 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-kubelet\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.063913 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cni-binary-copy\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064099 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-daemon-config\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064243 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064284 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-multus-certs\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064322 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-system-cni-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-cni-binary-copy\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064546 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cni-binary-copy\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064583 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-run-k8s-cni-cncf-io\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064606 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-cni-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064631 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-host-var-lib-kubelet\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064611 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cnibin\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064664 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-socket-dir-parent\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064714 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-multus-conf-dir\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064927 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.066651 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-os-release\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.066741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-cnibin\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.064969 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-os-release\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.086063 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.087868 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmnss\" (UniqueName: \"kubernetes.io/projected/be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235-kube-api-access-tmnss\") pod \"multus-jtx8z\" (UID: \"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\") " pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.094476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68w6k\" (UniqueName: \"kubernetes.io/projected/a2c74a28-06eb-4b6f-b540-d83b4f69f2a3-kube-api-access-68w6k\") pod \"multus-additional-cni-plugins-pbxln\" (UID: \"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\") " pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.102360 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jtx8z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.113330 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.114639 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pbxln" Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.118375 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe0cbbcc_9c4b_4ba8_87d2_f5dc13e57235.slice/crio-839cd12e8d19cd173369a73e2257401895c3954ed7a0043c8e6d42aa1d9d1899 WatchSource:0}: Error finding container 839cd12e8d19cd173369a73e2257401895c3954ed7a0043c8e6d42aa1d9d1899: Status 404 returned error can't find the container with id 839cd12e8d19cd173369a73e2257401895c3954ed7a0043c8e6d42aa1d9d1899 Dec 08 14:46:39 crc kubenswrapper[4894]: W1208 14:46:39.127725 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2c74a28_06eb_4b6f_b540_d83b4f69f2a3.slice/crio-3bbcea98400e502b67dbe7cb4fc0845a85899f586bb489088028b995758ee3fe WatchSource:0}: Error finding container 3bbcea98400e502b67dbe7cb4fc0845a85899f586bb489088028b995758ee3fe: Status 404 returned error can't find the container with id 3bbcea98400e502b67dbe7cb4fc0845a85899f586bb489088028b995758ee3fe Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.134238 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.164116 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.186117 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c25bk"] Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.186941 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.188881 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.189213 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.189414 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.189578 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.189797 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.190211 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.190664 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.192046 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.198602 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:39 crc kubenswrapper[4894]: E1208 14:46:39.198731 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.198789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:39 crc kubenswrapper[4894]: E1208 14:46:39.198863 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.203368 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.204607 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.206012 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.207430 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.210956 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.211696 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.211747 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.213947 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.214640 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.215682 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.216246 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.217400 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.218444 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.219384 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.220432 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.221130 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.222189 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.222806 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.223275 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.224929 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.227001 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.228086 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.228738 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.229199 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.230004 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.231104 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.233250 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.235169 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.237145 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.239201 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.240012 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.242060 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.242728 4894 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.242902 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.244968 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.247786 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.250000 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.250570 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.252051 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.253314 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.254014 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.254768 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.256188 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.256746 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.257903 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.259171 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.260065 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.261168 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.261851 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.262929 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.263945 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.265168 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.265809 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.266391 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.266055 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.268283 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.270180 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.271668 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.272378 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.287174 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.309473 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.310902 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.315159 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.322617 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.322684 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.322699 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"0f9e9b91eebc4c034ef3bcad6850b559d0eb835fe6bf8cc8b0a86cbd74c548f9"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.324278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerStarted","Data":"3bbcea98400e502b67dbe7cb4fc0845a85899f586bb489088028b995758ee3fe"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.329315 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerStarted","Data":"6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.329386 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerStarted","Data":"839cd12e8d19cd173369a73e2257401895c3954ed7a0043c8e6d42aa1d9d1899"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.336204 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rr2kq" event={"ID":"6f4dd416-8f0d-4efb-944d-1720665a8d2c","Type":"ContainerStarted","Data":"2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.336251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rr2kq" event={"ID":"6f4dd416-8f0d-4efb-944d-1720665a8d2c","Type":"ContainerStarted","Data":"cb83a2f0f235b219ca4bf2107489b832c49848c4e28197b37b622a54f54c0c79"} Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.340392 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.363174 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366811 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-netd\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366867 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-script-lib\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366901 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb212a28-36c5-440e-8965-986352c5d3ea-ovn-node-metrics-cert\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366924 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-log-socket\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366940 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6qpk\" (UniqueName: \"kubernetes.io/projected/fb212a28-36c5-440e-8965-986352c5d3ea-kube-api-access-k6qpk\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366963 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-ovn\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366980 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-kubelet\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.366995 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-node-log\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367014 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367041 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-bin\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367087 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367130 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-config\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367186 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-env-overrides\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367225 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-slash\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367257 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-systemd-units\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-var-lib-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367309 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-etc-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367355 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-netns\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.367377 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-systemd\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.382495 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.398622 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.417154 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.431995 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.447529 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.462293 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.467930 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468061 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-config\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468085 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-env-overrides\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468110 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-slash\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468136 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-systemd-units\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468156 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-var-lib-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468182 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-netns\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468219 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-systemd\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468239 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-etc-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468269 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-netd\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468317 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-script-lib\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468366 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb212a28-36c5-440e-8965-986352c5d3ea-ovn-node-metrics-cert\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468415 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-log-socket\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468447 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6qpk\" (UniqueName: \"kubernetes.io/projected/fb212a28-36c5-440e-8965-986352c5d3ea-kube-api-access-k6qpk\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468471 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-ovn\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468495 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-kubelet\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468528 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-node-log\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468551 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468583 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-bin\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468757 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-netd\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468879 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-slash\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468927 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-systemd-units\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.468965 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-var-lib-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469285 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-node-log\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469303 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469336 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-env-overrides\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-bin\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469340 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469391 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-systemd\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469413 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-netns\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469439 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-etc-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469557 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-openvswitch\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469884 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-ovn\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469926 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-kubelet\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469969 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-script-lib\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.469931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-log-socket\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.470127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-config\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.474385 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb212a28-36c5-440e-8965-986352c5d3ea-ovn-node-metrics-cert\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.483140 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.488577 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6qpk\" (UniqueName: \"kubernetes.io/projected/fb212a28-36c5-440e-8965-986352c5d3ea-kube-api-access-k6qpk\") pod \"ovnkube-node-c25bk\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.500424 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.514127 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.514877 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.549621 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.568587 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.591417 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.603976 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.620578 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.646434 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.664306 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.680904 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.695870 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.711905 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.735647 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.749291 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.769994 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.786921 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.799415 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.812683 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.828917 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.842445 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.856987 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.873133 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.897720 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.946380 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.947349 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:39 crc kubenswrapper[4894]: I1208 14:46:39.986016 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.013628 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.026586 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.046349 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.096609 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.106289 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.126561 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.146611 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.166161 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.186692 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.196284 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.196591 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.232222 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.260048 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.267358 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.296805 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.300343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.300411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.300428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.300610 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.308839 4894 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.309129 4894 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.310365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.310388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.310397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.310412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.310422 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.332131 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-z6s8q"] Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.332517 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.332913 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.336389 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.340052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.340091 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.340103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.340124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.340136 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.345721 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.345767 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95" exitCode=0 Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.345900 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.346132 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"06a57e2fb05e0b17ee1ffb63945245145b5f09f6faf6b17b6446ba18f3c515b0"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.358539 4894 generic.go:334] "Generic (PLEG): container finished" podID="a2c74a28-06eb-4b6f-b540-d83b4f69f2a3" containerID="357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d" exitCode=0 Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.360382 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerDied","Data":"357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d"} Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.365743 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.368602 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.370671 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.370731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.370748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.370771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.370786 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.383488 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.389238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.389284 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.389296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.389315 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.389327 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.389282 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.419283 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.427142 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.429145 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.432532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.432577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.432586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.432602 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.432612 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.450229 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.450713 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.451066 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.458898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.458933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.458943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.458967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.458977 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.479505 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnnfc\" (UniqueName: \"kubernetes.io/projected/60593d51-c757-4138-855d-6904f15385b4-kube-api-access-xnnfc\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.479801 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60593d51-c757-4138-855d-6904f15385b4-host\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.480098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/60593d51-c757-4138-855d-6904f15385b4-serviceca\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.496134 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.533433 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.565696 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.565809 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.565917 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.565930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.565953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.565966 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.581845 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnnfc\" (UniqueName: \"kubernetes.io/projected/60593d51-c757-4138-855d-6904f15385b4-kube-api-access-xnnfc\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.581910 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60593d51-c757-4138-855d-6904f15385b4-host\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.581943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/60593d51-c757-4138-855d-6904f15385b4-serviceca\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.582315 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60593d51-c757-4138-855d-6904f15385b4-host\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.583121 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/60593d51-c757-4138-855d-6904f15385b4-serviceca\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.593298 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.623547 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnnfc\" (UniqueName: \"kubernetes.io/projected/60593d51-c757-4138-855d-6904f15385b4-kube-api-access-xnnfc\") pod \"node-ca-z6s8q\" (UID: \"60593d51-c757-4138-855d-6904f15385b4\") " pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.654877 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.658043 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-z6s8q" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.668673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.668717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.668727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.668744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.668756 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: W1208 14:46:40.671643 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60593d51_c757_4138_855d_6904f15385b4.slice/crio-c7e3e93d82fe90c40bbcececfc039c6006ce1f5a57c1bf00686614120d9ccc18 WatchSource:0}: Error finding container c7e3e93d82fe90c40bbcececfc039c6006ce1f5a57c1bf00686614120d9ccc18: Status 404 returned error can't find the container with id c7e3e93d82fe90c40bbcececfc039c6006ce1f5a57c1bf00686614120d9ccc18 Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.697626 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.744012 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.772229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.772270 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.772280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.772298 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.772312 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.776297 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.817621 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.858196 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.875010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.875557 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.875574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.875636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.875652 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.886086 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.886266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.886336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.886370 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.886403 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:46:44.886346069 +0000 UTC m=+25.986352184 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.886444 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.886500 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:44.886480102 +0000 UTC m=+25.986486217 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.886524 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:44.886516603 +0000 UTC m=+25.986522718 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.895251 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.933408 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.978194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.978229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.978242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.978261 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.978273 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:40Z","lastTransitionTime":"2025-12-08T14:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.978475 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.987406 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:40 crc kubenswrapper[4894]: I1208 14:46:40.987482 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.987602 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.987635 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.987602 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.987711 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.987734 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.987651 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.988024 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:44.987789166 +0000 UTC m=+26.087795281 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:40 crc kubenswrapper[4894]: E1208 14:46:40.988064 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:44.988037703 +0000 UTC m=+26.088043998 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.021658 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.053478 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.080355 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.080407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.080419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.080438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.080453 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.092291 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.135554 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.180917 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.182477 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.182510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.182521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.182538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.182549 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.196696 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:41 crc kubenswrapper[4894]: E1208 14:46:41.196846 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.196697 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:41 crc kubenswrapper[4894]: E1208 14:46:41.197062 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.214101 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.252880 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.254538 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.256915 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.271842 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.285072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.285114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.285123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.285140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.285152 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.312873 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.353727 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.362976 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-z6s8q" event={"ID":"60593d51-c757-4138-855d-6904f15385b4","Type":"ContainerStarted","Data":"2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.363053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-z6s8q" event={"ID":"60593d51-c757-4138-855d-6904f15385b4","Type":"ContainerStarted","Data":"c7e3e93d82fe90c40bbcececfc039c6006ce1f5a57c1bf00686614120d9ccc18"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.364720 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.369862 4894 generic.go:334] "Generic (PLEG): container finished" podID="a2c74a28-06eb-4b6f-b540-d83b4f69f2a3" containerID="8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996" exitCode=0 Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.369956 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerDied","Data":"8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.377310 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.377349 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.377365 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.377378 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.377390 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.377402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.396189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.396229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.396238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.396256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.396265 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.398004 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: E1208 14:46:41.409984 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.456197 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.496176 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.499189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.499229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.499238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.499255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.499264 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.537628 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.578736 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.602114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.602168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.602180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.602200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.602212 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.618937 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.655025 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.696413 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.704372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.704417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.704429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.704447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.704460 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.741944 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.777134 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.807393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.807445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.807458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.807481 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.807492 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.822422 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.859987 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.900183 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.909035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.909073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.909082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.909099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.909108 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:41Z","lastTransitionTime":"2025-12-08T14:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.936866 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:41 crc kubenswrapper[4894]: I1208 14:46:41.973579 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:41Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.011144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.011195 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.011207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.011225 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.011240 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.015064 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.057963 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.096559 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.114304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.114349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.114362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.114380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.114392 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.138045 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.189687 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.196144 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:42 crc kubenswrapper[4894]: E1208 14:46:42.196400 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.217530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.217596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.217610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.217632 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.217700 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.227194 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.320982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.321034 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.321044 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.321062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.321073 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.382761 4894 generic.go:334] "Generic (PLEG): container finished" podID="a2c74a28-06eb-4b6f-b540-d83b4f69f2a3" containerID="852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2" exitCode=0 Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.383556 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerDied","Data":"852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.419128 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.425266 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.425419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.425447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.425476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.425495 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.445857 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.462225 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.477491 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.491937 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.505281 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.528499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.528546 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.528555 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.528571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.528582 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.529759 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.545529 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.574993 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.615846 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.631574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.631632 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.631644 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.631660 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.631725 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.653212 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.694827 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.735079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.735119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.735131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.735147 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.735130 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.735158 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.775497 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.814389 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:42Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.838907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.838964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.838977 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.838999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.839013 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.942282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.942328 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.942338 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.942354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:42 crc kubenswrapper[4894]: I1208 14:46:42.942365 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:42Z","lastTransitionTime":"2025-12-08T14:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.045203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.045255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.045271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.045293 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.045306 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.147557 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.147606 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.147617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.147633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.147645 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.196500 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.196545 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:43 crc kubenswrapper[4894]: E1208 14:46:43.196732 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:43 crc kubenswrapper[4894]: E1208 14:46:43.196915 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.251743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.251839 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.251877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.251907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.251926 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.356742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.356865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.356899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.356931 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.356955 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.391744 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.394660 4894 generic.go:334] "Generic (PLEG): container finished" podID="a2c74a28-06eb-4b6f-b540-d83b4f69f2a3" containerID="368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334" exitCode=0 Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.394708 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerDied","Data":"368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.416136 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.432727 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.451269 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.460290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.460572 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.460583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.460601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.460613 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.464240 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.476784 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.491233 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.510025 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.525327 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.538401 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.552436 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.563261 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.563308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.563324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.563343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.563355 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.564088 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.576390 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.592207 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.604068 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.684635 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.684679 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.684688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.684704 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.684716 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.688343 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:43Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.787410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.787438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.787447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.787462 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.787470 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.889263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.889288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.889297 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.889310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.889319 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.991927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.991976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.991992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.992017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:43 crc kubenswrapper[4894]: I1208 14:46:43.992034 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:43Z","lastTransitionTime":"2025-12-08T14:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.093905 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.093947 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.093955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.093970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.093980 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.195945 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:44 crc kubenswrapper[4894]: E1208 14:46:44.196076 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.197700 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.197733 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.197744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.197757 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.197769 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.299841 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.299880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.299889 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.299904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.299912 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.407052 4894 generic.go:334] "Generic (PLEG): container finished" podID="a2c74a28-06eb-4b6f-b540-d83b4f69f2a3" containerID="7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0" exitCode=0 Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.407127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerDied","Data":"7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.407890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.407927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.407939 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.407956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.407966 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.425853 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.444834 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.465954 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.479021 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.493647 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.507340 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.512479 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.512563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.512578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.512596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.512608 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.529888 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.543679 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.568487 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.583071 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.594568 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.608645 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.615783 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.615874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.615890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.615912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.615925 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.628424 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.643293 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.656053 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:44Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.718595 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.718635 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.718644 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.718659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.718670 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.821446 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.821489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.821497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.821514 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.821523 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.924199 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.924242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.924252 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.924270 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.924284 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:44Z","lastTransitionTime":"2025-12-08T14:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.932884 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.933060 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:44 crc kubenswrapper[4894]: E1208 14:46:44.933140 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:46:52.933099922 +0000 UTC m=+34.033106167 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:46:44 crc kubenswrapper[4894]: E1208 14:46:44.933213 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:44 crc kubenswrapper[4894]: E1208 14:46:44.933306 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:52.933279207 +0000 UTC m=+34.033285332 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:44 crc kubenswrapper[4894]: E1208 14:46:44.933339 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:44 crc kubenswrapper[4894]: E1208 14:46:44.933392 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:52.93337992 +0000 UTC m=+34.033386035 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:44 crc kubenswrapper[4894]: I1208 14:46:44.933213 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.027419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.027474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.027490 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.027513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.027529 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.034091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.034163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034367 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034395 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034410 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034474 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:53.034452217 +0000 UTC m=+34.134458332 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034484 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034506 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034517 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.034555 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:46:53.03454246 +0000 UTC m=+34.134548585 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.129870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.129911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.129919 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.129937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.129950 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.196326 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.196326 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.196649 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:45 crc kubenswrapper[4894]: E1208 14:46:45.196761 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.232182 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.232258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.232283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.232314 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.232336 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.335247 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.335317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.335326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.335343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.335353 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.415293 4894 generic.go:334] "Generic (PLEG): container finished" podID="a2c74a28-06eb-4b6f-b540-d83b4f69f2a3" containerID="51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51" exitCode=0 Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.415335 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerDied","Data":"51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.438542 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.439858 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.439890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.439898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.439915 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.439925 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.462518 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.478974 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.496008 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.510483 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.524424 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.542771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.543202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.543212 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.543229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.543241 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.545315 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.563523 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.581633 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.599439 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.610442 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.620865 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.629720 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.640506 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.646100 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.646136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.646148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.646180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.646193 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.651633 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:45Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.748408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.748437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.748445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.748463 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.748476 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.850221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.850285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.850298 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.850323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.850362 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.953682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.953786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.953842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.953877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:45 crc kubenswrapper[4894]: I1208 14:46:45.953901 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:45Z","lastTransitionTime":"2025-12-08T14:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.056363 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.056421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.056435 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.056458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.056476 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.159628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.159676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.159686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.159707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.159724 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.196688 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:46 crc kubenswrapper[4894]: E1208 14:46:46.196867 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.262645 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.262702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.262713 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.262733 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.262745 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.365999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.366053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.366065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.366085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.366097 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.421150 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.421379 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.421418 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.424949 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" event={"ID":"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3","Type":"ContainerStarted","Data":"5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.434860 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.445725 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.447622 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.448305 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.462648 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.468700 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.468734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.468744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.468760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.468771 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.479324 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.492182 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.506239 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.516679 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.527081 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.540222 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.553444 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.567249 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.571221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.571265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.571275 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.571296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.571308 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.579382 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.598729 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.622161 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.637345 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.658440 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.674563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.674623 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.674637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.674659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.674670 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.678425 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.693401 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.702346 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.713551 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.728908 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.762212 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.776319 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.777228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.777267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.777278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.777294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.777305 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.793283 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.811057 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.831576 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.848400 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.864839 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.878096 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.879607 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.879656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.879669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.879686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.879698 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.891163 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:46Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.983163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.983233 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.983253 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.983280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:46 crc kubenswrapper[4894]: I1208 14:46:46.983298 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:46Z","lastTransitionTime":"2025-12-08T14:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.086251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.086308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.086322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.086340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.086350 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.189478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.189518 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.189528 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.189544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.189553 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.196032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.196058 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:47 crc kubenswrapper[4894]: E1208 14:46:47.196137 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:47 crc kubenswrapper[4894]: E1208 14:46:47.196221 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.292358 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.292402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.292411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.292426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.292434 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.394943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.394985 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.394996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.395010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.395019 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.427302 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.496803 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.496861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.496871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.496887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.496896 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.600046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.600101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.600118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.600140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.600154 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.702602 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.702634 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.702643 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.702656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.702665 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.804765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.804798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.804807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.804839 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.804850 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.906873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.906903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.906912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.906926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:47 crc kubenswrapper[4894]: I1208 14:46:47.906934 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:47Z","lastTransitionTime":"2025-12-08T14:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.008989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.009021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.009030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.009046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.009056 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.112160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.112201 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.112210 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.112226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.112236 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.196340 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:48 crc kubenswrapper[4894]: E1208 14:46:48.196570 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.216149 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.216203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.216221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.216246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.216264 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.318458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.318530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.318544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.318560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.318574 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.420842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.420870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.420878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.420892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.420902 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.431555 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/0.log" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.434865 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265" exitCode=1 Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.434909 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.435462 4894 scope.go:117] "RemoveContainer" containerID="d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.447810 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.471027 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.492138 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.512635 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.525449 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.525486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.525495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.525512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.525524 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.536472 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.552397 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.572637 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:48Z\\\",\\\"message\\\":\\\"ontroller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 14:46:48.163649 6169 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1208 14:46:48.163687 6169 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1208 14:46:48.163726 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:48.163779 6169 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:48.163780 6169 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 14:46:48.163808 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:48.163858 6169 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1208 14:46:48.163909 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1208 14:46:48.163926 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1208 14:46:48.163946 6169 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1208 14:46:48.163957 6169 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 14:46:48.163970 6169 factory.go:656] Stopping watch factory\\\\nI1208 14:46:48.163978 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1208 14:46:48.163991 6169 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:48.163999 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1208 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.591921 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.610538 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.625347 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.628431 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.628481 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.628494 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.628516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.628532 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.646369 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.663658 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.681490 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.698086 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.710154 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:48Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.732332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.732366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.732378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.732399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.732413 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.834871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.834900 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.834907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.834920 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.834941 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.937421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.937492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.937503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.937519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:48 crc kubenswrapper[4894]: I1208 14:46:48.937530 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:48Z","lastTransitionTime":"2025-12-08T14:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.040521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.040566 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.040578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.040599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.040615 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.143148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.143233 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.143249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.143276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.143293 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.196119 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.196196 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:49 crc kubenswrapper[4894]: E1208 14:46:49.196338 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:49 crc kubenswrapper[4894]: E1208 14:46:49.196475 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.211746 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.227368 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.246003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.246203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.246230 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.246252 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.246265 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.246393 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.261903 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.280964 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.310231 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.327244 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.350668 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:48Z\\\",\\\"message\\\":\\\"ontroller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 14:46:48.163649 6169 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1208 14:46:48.163687 6169 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1208 14:46:48.163726 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:48.163779 6169 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:48.163780 6169 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 14:46:48.163808 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:48.163858 6169 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1208 14:46:48.163909 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1208 14:46:48.163926 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1208 14:46:48.163946 6169 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1208 14:46:48.163957 6169 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 14:46:48.163970 6169 factory.go:656] Stopping watch factory\\\\nI1208 14:46:48.163978 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1208 14:46:48.163991 6169 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:48.163999 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1208 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.364028 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.386072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.386115 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.386125 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.386141 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.386150 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.390144 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.404609 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.415291 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.427176 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.440373 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.440551 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/0.log" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.444460 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.444578 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.454276 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.469850 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.482991 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.488992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.489040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.489054 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.489074 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.489087 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.494528 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.503302 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.518829 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.537664 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.562940 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.581734 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.592107 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.592163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.592175 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.592190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.592201 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.597439 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.611468 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.623290 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.635058 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.645847 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.660151 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.689421 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:48Z\\\",\\\"message\\\":\\\"ontroller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 14:46:48.163649 6169 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1208 14:46:48.163687 6169 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1208 14:46:48.163726 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:48.163779 6169 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:48.163780 6169 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 14:46:48.163808 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:48.163858 6169 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1208 14:46:48.163909 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1208 14:46:48.163926 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1208 14:46:48.163946 6169 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1208 14:46:48.163957 6169 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 14:46:48.163970 6169 factory.go:656] Stopping watch factory\\\\nI1208 14:46:48.163978 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1208 14:46:48.163991 6169 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:48.163999 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1208 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:49Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.695067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.695152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.695162 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.695182 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.695194 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.801691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.801739 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.801751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.801770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.801781 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.904233 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.904279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.904288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.904305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:49 crc kubenswrapper[4894]: I1208 14:46:49.904316 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:49Z","lastTransitionTime":"2025-12-08T14:46:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.007183 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.007244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.007256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.007283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.007298 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.110080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.110145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.110155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.110171 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.110183 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.196171 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.196361 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.213170 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.213227 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.213243 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.213265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.213281 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.317314 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.317418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.317437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.317483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.317507 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.420918 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.420991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.421010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.421037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.421057 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.451111 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/1.log" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.452114 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/0.log" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.455734 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775" exitCode=1 Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.455802 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.455880 4894 scope.go:117] "RemoveContainer" containerID="d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.456673 4894 scope.go:117] "RemoveContainer" containerID="875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775" Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.456909 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.472147 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.491506 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:48Z\\\",\\\"message\\\":\\\"ontroller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 14:46:48.163649 6169 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1208 14:46:48.163687 6169 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1208 14:46:48.163726 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:48.163779 6169 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:48.163780 6169 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 14:46:48.163808 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:48.163858 6169 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1208 14:46:48.163909 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1208 14:46:48.163926 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1208 14:46:48.163946 6169 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1208 14:46:48.163957 6169 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 14:46:48.163970 6169 factory.go:656] Stopping watch factory\\\\nI1208 14:46:48.163978 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1208 14:46:48.163991 6169 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:48.163999 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1208 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.511727 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.523179 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.523233 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.523249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.523271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.523288 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.525780 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.540771 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.553788 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.571488 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.585321 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.598172 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.612088 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.625751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.625895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.625925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.625962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.625988 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.629431 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.649649 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.682766 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.704189 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.728248 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.730436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.730487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.730507 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.730534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.730561 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.732335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.732397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.732414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.732443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.732462 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.754668 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.760617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.760662 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.760675 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.760697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.760712 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.781017 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.787340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.787412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.787436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.787468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.787493 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.813992 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.819147 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.819196 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.819209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.819230 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.819245 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.834153 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.838364 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.838415 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.838428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.838451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.838467 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.855074 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:50Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:50 crc kubenswrapper[4894]: E1208 14:46:50.855247 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.857690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.857740 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.857755 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.857779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.857797 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.961272 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.961332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.961343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.961367 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:50 crc kubenswrapper[4894]: I1208 14:46:50.961384 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:50Z","lastTransitionTime":"2025-12-08T14:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.037629 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.063495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.063987 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.064160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.064296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.064409 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.066662 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.082765 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.103826 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.120499 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.145188 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:48Z\\\",\\\"message\\\":\\\"ontroller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 14:46:48.163649 6169 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1208 14:46:48.163687 6169 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1208 14:46:48.163726 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:48.163779 6169 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:48.163780 6169 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 14:46:48.163808 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:48.163858 6169 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1208 14:46:48.163909 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1208 14:46:48.163926 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1208 14:46:48.163946 6169 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1208 14:46:48.163957 6169 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 14:46:48.163970 6169 factory.go:656] Stopping watch factory\\\\nI1208 14:46:48.163978 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1208 14:46:48.163991 6169 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:48.163999 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1208 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.164360 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5"] Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.164855 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.165373 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.167027 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.167063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.167073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.167088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.167102 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.167563 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.167846 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.180201 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.194480 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.198055 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.198087 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:51 crc kubenswrapper[4894]: E1208 14:46:51.198213 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:51 crc kubenswrapper[4894]: E1208 14:46:51.198325 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.210014 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.226874 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.240736 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.255131 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.265029 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.269077 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.269154 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.269163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.269180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.269190 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.278671 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.292913 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.304324 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-994vg\" (UniqueName: \"kubernetes.io/projected/57e7a706-dfa9-4c32-a7bd-478d8faa771b-kube-api-access-994vg\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.304632 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/57e7a706-dfa9-4c32-a7bd-478d8faa771b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.304768 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/57e7a706-dfa9-4c32-a7bd-478d8faa771b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.304917 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/57e7a706-dfa9-4c32-a7bd-478d8faa771b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.311944 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.327124 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.342053 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.355126 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.375843 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.375912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.375923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.375948 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.375960 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.376119 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d7f31e885bbd0b4efccc3faf18fd6760d0fc03cf8fc2ef8846ea00b42e6265\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:48Z\\\",\\\"message\\\":\\\"ontroller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1208 14:46:48.163649 6169 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1208 14:46:48.163687 6169 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1208 14:46:48.163726 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:48.163779 6169 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:48.163780 6169 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1208 14:46:48.163808 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:48.163858 6169 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1208 14:46:48.163909 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1208 14:46:48.163926 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1208 14:46:48.163946 6169 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1208 14:46:48.163957 6169 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1208 14:46:48.163970 6169 factory.go:656] Stopping watch factory\\\\nI1208 14:46:48.163978 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1208 14:46:48.163991 6169 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:48.163999 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1208 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.392426 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.406422 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/57e7a706-dfa9-4c32-a7bd-478d8faa771b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.406482 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-994vg\" (UniqueName: \"kubernetes.io/projected/57e7a706-dfa9-4c32-a7bd-478d8faa771b-kube-api-access-994vg\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.406513 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/57e7a706-dfa9-4c32-a7bd-478d8faa771b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.406538 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/57e7a706-dfa9-4c32-a7bd-478d8faa771b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.407298 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/57e7a706-dfa9-4c32-a7bd-478d8faa771b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.407536 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/57e7a706-dfa9-4c32-a7bd-478d8faa771b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.413438 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.414677 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/57e7a706-dfa9-4c32-a7bd-478d8faa771b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.427116 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-994vg\" (UniqueName: \"kubernetes.io/projected/57e7a706-dfa9-4c32-a7bd-478d8faa771b-kube-api-access-994vg\") pod \"ovnkube-control-plane-749d76644c-zgtl5\" (UID: \"57e7a706-dfa9-4c32-a7bd-478d8faa771b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.432322 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.449026 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.464714 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.465167 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/1.log" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.471647 4894 scope.go:117] "RemoveContainer" containerID="875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775" Dec 08 14:46:51 crc kubenswrapper[4894]: E1208 14:46:51.471949 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.478899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.478928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.478937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.478953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.478963 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.480251 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.481559 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.494734 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: W1208 14:46:51.500361 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57e7a706_dfa9_4c32_a7bd_478d8faa771b.slice/crio-3500c704170fe07d8dd10fa3ff448276b60495d4751c18a9621bb439b14cb8fc WatchSource:0}: Error finding container 3500c704170fe07d8dd10fa3ff448276b60495d4751c18a9621bb439b14cb8fc: Status 404 returned error can't find the container with id 3500c704170fe07d8dd10fa3ff448276b60495d4751c18a9621bb439b14cb8fc Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.511339 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.527516 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.544128 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.565745 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.581785 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.582584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.582618 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.582634 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.582655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.582667 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.594465 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.607207 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.627135 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.641918 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.655569 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.668150 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.679968 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.685497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.685527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.685536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.685551 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.685562 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.694968 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.707162 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.719284 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.732160 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.746594 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.767978 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.780040 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.789038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.789081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.789094 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.789113 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.789125 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.794696 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:51Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.893206 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.893705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.893731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.893771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.893801 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.997345 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.997397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.997407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.997428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:51 crc kubenswrapper[4894]: I1208 14:46:51.997438 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:51Z","lastTransitionTime":"2025-12-08T14:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.099945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.099989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.099998 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.100015 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.100026 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.196647 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:52 crc kubenswrapper[4894]: E1208 14:46:52.196874 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.202694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.202741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.202759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.202782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.202797 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.267992 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-z2zz6"] Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.268649 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:52 crc kubenswrapper[4894]: E1208 14:46:52.268738 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.291719 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.305728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.305773 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.305785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.305802 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.305850 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.308351 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.321549 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.341969 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.357193 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.371480 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.382863 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.398448 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.408411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.408450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.408459 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.408474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.408485 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.416875 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.416947 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmvwm\" (UniqueName: \"kubernetes.io/projected/49a05fc4-1361-4ebc-891a-e2b49df28ffa-kube-api-access-fmvwm\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.421340 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.435689 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.450376 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.466498 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.475043 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" event={"ID":"57e7a706-dfa9-4c32-a7bd-478d8faa771b","Type":"ContainerStarted","Data":"8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.475091 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" event={"ID":"57e7a706-dfa9-4c32-a7bd-478d8faa771b","Type":"ContainerStarted","Data":"a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.475106 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" event={"ID":"57e7a706-dfa9-4c32-a7bd-478d8faa771b","Type":"ContainerStarted","Data":"3500c704170fe07d8dd10fa3ff448276b60495d4751c18a9621bb439b14cb8fc"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.483944 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.500154 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.510759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.510788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.510862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.510879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.510889 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.517276 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmvwm\" (UniqueName: \"kubernetes.io/projected/49a05fc4-1361-4ebc-891a-e2b49df28ffa-kube-api-access-fmvwm\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.517321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:52 crc kubenswrapper[4894]: E1208 14:46:52.517418 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:52 crc kubenswrapper[4894]: E1208 14:46:52.517463 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:46:53.017450376 +0000 UTC m=+34.117456491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.520119 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.545707 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmvwm\" (UniqueName: \"kubernetes.io/projected/49a05fc4-1361-4ebc-891a-e2b49df28ffa-kube-api-access-fmvwm\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.560710 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.581101 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.609451 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.613011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.613048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.613057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.613072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.613082 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.621449 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.634732 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.646959 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.657052 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.667576 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.687789 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.700760 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.715121 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.716499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.716631 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.716710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.716792 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.716808 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.726366 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.737644 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.750276 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.764528 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.775651 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.785907 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.798432 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.811725 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:52Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.819997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.820051 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.820065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.820086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.820100 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.922673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.922720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.922748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.922765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:52 crc kubenswrapper[4894]: I1208 14:46:52.922776 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:52Z","lastTransitionTime":"2025-12-08T14:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.022597 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.022728 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:47:09.022702176 +0000 UTC m=+50.122708301 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.022805 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.022908 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.022960 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.022993 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.023051 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:09.023039696 +0000 UTC m=+50.123045811 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.023061 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.023150 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.023209 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:09.02317927 +0000 UTC m=+50.123185415 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.023239 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:46:54.023225671 +0000 UTC m=+35.123231826 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.025777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.025898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.025918 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.025945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.025964 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.124069 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.124130 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124286 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124323 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124334 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124492 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124549 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124573 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124687 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:09.124655889 +0000 UTC m=+50.224662044 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.124795 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:09.124779102 +0000 UTC m=+50.224785217 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.128859 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.128911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.128923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.128939 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.128978 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.196972 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.197269 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.197433 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:53 crc kubenswrapper[4894]: E1208 14:46:53.197603 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.231772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.231861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.231872 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.231892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.231905 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.335425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.335486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.335502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.335529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.335550 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.437994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.438032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.438040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.438055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.438064 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.540040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.540074 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.540083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.540097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.540106 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.642737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.642778 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.642790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.642809 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.642839 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.745651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.745690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.745702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.745718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.745730 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.849217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.849259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.849269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.849286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.849298 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.952470 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.952538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.952555 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.952580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:53 crc kubenswrapper[4894]: I1208 14:46:53.952596 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:53Z","lastTransitionTime":"2025-12-08T14:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.034357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:54 crc kubenswrapper[4894]: E1208 14:46:54.034490 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:54 crc kubenswrapper[4894]: E1208 14:46:54.034553 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:46:56.034534046 +0000 UTC m=+37.134540211 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.055267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.055336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.055351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.055374 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.055389 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.159012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.159080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.159101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.159127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.159146 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.196750 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.196791 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:54 crc kubenswrapper[4894]: E1208 14:46:54.196929 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:54 crc kubenswrapper[4894]: E1208 14:46:54.197067 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.261667 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.261709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.261722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.261806 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.261847 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.363647 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.363712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.363730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.363747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.363759 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.466313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.466390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.466407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.466423 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.466434 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.570485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.570550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.570567 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.570594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.570611 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.673789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.673875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.673889 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.673909 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.673922 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.777064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.777126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.777134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.777150 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.777180 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.882072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.882140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.882154 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.882178 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.882194 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.985400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.985445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.985457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.985475 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:54 crc kubenswrapper[4894]: I1208 14:46:54.985489 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:54Z","lastTransitionTime":"2025-12-08T14:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.088177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.088234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.088246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.088269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.088283 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.191030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.191113 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.191138 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.191173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.191198 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.196265 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.196264 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:55 crc kubenswrapper[4894]: E1208 14:46:55.196421 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:55 crc kubenswrapper[4894]: E1208 14:46:55.196522 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.295471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.295525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.295540 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.295561 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.295575 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.398842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.398876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.398886 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.398899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.398908 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.500955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.501003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.501013 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.501028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.501039 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.603620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.603655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.603666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.603683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.603698 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.705785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.705853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.705864 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.705880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.705889 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.808834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.808875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.808886 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.808901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.808912 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.911580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.911624 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.911634 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.911650 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:55 crc kubenswrapper[4894]: I1208 14:46:55.911660 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:55Z","lastTransitionTime":"2025-12-08T14:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.014453 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.014715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.014727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.014748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.014757 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.056146 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:56 crc kubenswrapper[4894]: E1208 14:46:56.056289 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:56 crc kubenswrapper[4894]: E1208 14:46:56.056341 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:47:00.056326452 +0000 UTC m=+41.156332567 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.116775 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.116846 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.116861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.116881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.116896 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.195906 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.195976 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:56 crc kubenswrapper[4894]: E1208 14:46:56.196086 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:56 crc kubenswrapper[4894]: E1208 14:46:56.196164 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.219304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.219349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.219360 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.219379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.219392 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.322130 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.322174 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.322185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.322201 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.322210 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.424745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.424784 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.424800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.424833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.424844 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.531687 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.531749 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.531766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.531790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.531807 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.635515 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.635560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.635570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.635586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.635597 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.739042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.739177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.739202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.739237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.739262 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.843066 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.843140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.843155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.843183 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.843199 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.947213 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.947290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.947304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.947326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:56 crc kubenswrapper[4894]: I1208 14:46:56.947340 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:56Z","lastTransitionTime":"2025-12-08T14:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.049520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.049557 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.049566 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.049579 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.049591 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.151693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.151765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.151779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.151798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.151810 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.196439 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:57 crc kubenswrapper[4894]: E1208 14:46:57.196582 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.196702 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:57 crc kubenswrapper[4894]: E1208 14:46:57.196958 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.255195 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.255255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.255268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.255292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.255310 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.358213 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.358282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.358302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.358329 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.358347 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.462151 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.462240 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.462265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.462337 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.462364 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.565957 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.566028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.566061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.566090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.566110 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.672041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.672091 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.672103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.672121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.672135 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.774067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.774099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.774107 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.774120 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.774129 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.876375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.876417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.876428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.876445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.876455 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.979410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.979461 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.979474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.979492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:57 crc kubenswrapper[4894]: I1208 14:46:57.979506 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:57Z","lastTransitionTime":"2025-12-08T14:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.082322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.082379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.082402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.082430 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.082451 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.184714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.184750 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.184758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.184772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.184781 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.196045 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:46:58 crc kubenswrapper[4894]: E1208 14:46:58.196187 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.196070 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:46:58 crc kubenswrapper[4894]: E1208 14:46:58.196750 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.287361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.287391 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.287399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.287415 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.287423 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.388992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.389033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.389047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.389066 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.389078 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.491742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.491795 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.491851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.491868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.491879 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.594411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.594456 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.594467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.594486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.594498 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.723805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.723929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.723949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.723976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.723995 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.826515 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.826597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.826622 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.826654 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.826678 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.929505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.930137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.930162 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.930191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:58 crc kubenswrapper[4894]: I1208 14:46:58.930208 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:58Z","lastTransitionTime":"2025-12-08T14:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.032590 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.032637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.032649 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.032699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.032712 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.135481 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.135525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.135536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.135553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.135565 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.196423 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.196480 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:46:59 crc kubenswrapper[4894]: E1208 14:46:59.196565 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:46:59 crc kubenswrapper[4894]: E1208 14:46:59.197857 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.210871 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.222168 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.236522 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.238343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.238383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.238395 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.238414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.238426 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.245600 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.263680 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.275080 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.287298 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.305136 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.326393 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.341246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.341310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.341327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.341352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.341370 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.345315 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.357908 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.369370 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.379977 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.389770 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.405795 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.419951 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.435355 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:46:59Z is after 2025-08-24T17:21:41Z" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.444289 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.444382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.444405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.444430 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.444448 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.547970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.548030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.548047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.548118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.548179 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.651382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.651419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.651448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.651466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.651475 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.754565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.754597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.754605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.754620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.754629 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.858006 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.858471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.858488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.858511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.858526 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.961511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.961589 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.961601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.961620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:46:59 crc kubenswrapper[4894]: I1208 14:46:59.961633 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:46:59Z","lastTransitionTime":"2025-12-08T14:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.063573 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.063656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.063668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.063688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.063701 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.103151 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:00 crc kubenswrapper[4894]: E1208 14:47:00.103345 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:00 crc kubenswrapper[4894]: E1208 14:47:00.103470 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:47:08.103434185 +0000 UTC m=+49.203440300 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.165883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.165916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.165943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.165956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.165964 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.196320 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:00 crc kubenswrapper[4894]: E1208 14:47:00.196448 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.196693 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:00 crc kubenswrapper[4894]: E1208 14:47:00.197030 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.268613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.268718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.268731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.268747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.268757 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.370960 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.371200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.371323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.371419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.371515 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.474637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.474673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.474681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.474693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.474701 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.577592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.577660 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.577673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.577690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.577704 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.683276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.684441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.684580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.684726 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.684933 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.788167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.788646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.788862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.789062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.789303 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.893476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.893927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.894352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.895009 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.895107 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.998576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.998645 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.998662 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.998725 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:00 crc kubenswrapper[4894]: I1208 14:47:00.998742 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:00Z","lastTransitionTime":"2025-12-08T14:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.102356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.102420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.102441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.102473 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.102496 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.120601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.120719 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.120744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.120774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.120800 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.142423 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:01Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.147087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.147118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.147126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.147141 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.147151 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.161209 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:01Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.165897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.165967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.165983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.166015 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.166033 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.183494 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:01Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.188340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.188418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.188440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.188466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.188483 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.196537 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.196604 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.196666 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.196751 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.204584 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:01Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.208853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.208891 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.208904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.208920 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.208935 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.223946 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:01Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:01 crc kubenswrapper[4894]: E1208 14:47:01.224089 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.225769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.225836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.225848 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.225868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.225915 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.328286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.328317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.328326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.328339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.328347 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.430925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.430973 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.430986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.431008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.431022 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.533836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.533879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.533892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.533911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.533924 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.636964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.637012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.637022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.637036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.637047 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.740325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.740380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.740399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.740420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.740431 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.842435 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.842542 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.842553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.842568 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.842580 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.945406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.945445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.945455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.945471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:01 crc kubenswrapper[4894]: I1208 14:47:01.945484 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:01Z","lastTransitionTime":"2025-12-08T14:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.047438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.047496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.047513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.047547 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.047572 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.150083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.150140 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.150156 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.150180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.150197 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.195837 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.195951 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:02 crc kubenswrapper[4894]: E1208 14:47:02.195975 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:02 crc kubenswrapper[4894]: E1208 14:47:02.196172 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.197201 4894 scope.go:117] "RemoveContainer" containerID="875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.252274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.252310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.252318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.252332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.252342 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.355235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.355519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.355585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.355646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.355703 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.459106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.459143 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.459152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.459166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.459175 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.507988 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/1.log" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.510192 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.510316 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.523791 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.539168 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.550401 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.562135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.562178 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.562191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.562209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.562222 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.564012 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.578160 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.598603 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.616232 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.648695 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.664454 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.664497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.664512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.664529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.664541 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.665937 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.689138 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.703845 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.713842 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.727086 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.745896 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.759252 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.766711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.766743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.766753 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.766772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.766783 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.770719 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.783566 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.869389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.869427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.869439 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.869455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.869465 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.971655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.971682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.971691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.971705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:02 crc kubenswrapper[4894]: I1208 14:47:02.971714 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:02Z","lastTransitionTime":"2025-12-08T14:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.074384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.074425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.074434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.074448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.074459 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.176866 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.176911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.176922 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.176939 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.176947 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.196373 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:03 crc kubenswrapper[4894]: E1208 14:47:03.196462 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.196377 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:03 crc kubenswrapper[4894]: E1208 14:47:03.196543 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.279604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.279666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.279675 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.279692 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.279702 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.382753 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.382805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.382887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.382924 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.382947 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.487690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.487724 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.487734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.487756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.487767 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.516246 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/2.log" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.517173 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/1.log" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.520123 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be" exitCode=1 Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.520162 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.520220 4894 scope.go:117] "RemoveContainer" containerID="875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.520850 4894 scope.go:117] "RemoveContainer" containerID="26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be" Dec 08 14:47:03 crc kubenswrapper[4894]: E1208 14:47:03.521023 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.543010 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.562560 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.589690 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.591073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.591142 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.591161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.591191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.591213 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.605125 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.625376 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.636515 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.649300 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.668164 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.687069 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://875167a29af84d2cbfc8dfee41df8542c49203a5be7c5414726cd6e2da249775\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:46:49Z\\\",\\\"message\\\":\\\"BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325579 6294 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.325621 6294 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1208 14:46:49.325698 6294 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1208 14:46:49.326139 6294 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1208 14:46:49.326176 6294 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1208 14:46:49.326181 6294 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1208 14:46:49.326200 6294 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1208 14:46:49.326206 6294 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1208 14:46:49.326232 6294 factory.go:656] Stopping watch factory\\\\nI1208 14:46:49.326246 6294 ovnkube.go:599] Stopped ovnkube\\\\nI1208 14:46:49.326247 6294 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1208 14:46:49.326271 6294 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1208 14:46:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.694166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.694208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.694225 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.694243 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.694254 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.707844 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.729359 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.743144 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.756775 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.770171 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.783318 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.796646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.796690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.796702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.796720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.796731 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.798465 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.809645 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:03Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.899645 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.899773 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.899811 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.900084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:03 crc kubenswrapper[4894]: I1208 14:47:03.900102 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:03Z","lastTransitionTime":"2025-12-08T14:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.003067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.003109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.003119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.003134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.003145 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.105322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.105358 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.105370 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.105386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.105398 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.196521 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.196562 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:04 crc kubenswrapper[4894]: E1208 14:47:04.196640 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:04 crc kubenswrapper[4894]: E1208 14:47:04.196695 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.207763 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.207795 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.207806 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.207832 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.207840 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.310020 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.310071 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.310080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.310095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.310104 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.412997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.413033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.413046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.413062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.413074 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.516094 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.516148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.516166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.516188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.516207 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.524590 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/2.log" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.618417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.618457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.618469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.618483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.618492 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.721570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.721635 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.721653 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.721677 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.721698 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.825513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.825568 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.825576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.825592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.825603 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.929326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.929384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.929400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.929427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:04 crc kubenswrapper[4894]: I1208 14:47:04.929444 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:04Z","lastTransitionTime":"2025-12-08T14:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.031899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.031952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.031967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.031986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.031997 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.135294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.135336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.135348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.135365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.135376 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.196194 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.196245 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:05 crc kubenswrapper[4894]: E1208 14:47:05.196333 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:05 crc kubenswrapper[4894]: E1208 14:47:05.196383 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.237522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.237575 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.237586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.237604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.237615 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.339989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.340020 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.340028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.340041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.340050 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.442625 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.442669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.442685 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.442707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.442724 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.545010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.545078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.545095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.545124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.545142 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.647941 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.647997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.648007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.648022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.648032 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.750340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.750380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.750390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.750405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.750416 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.853647 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.853705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.853714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.853728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.853737 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.955901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.955941 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.955952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.955966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:05 crc kubenswrapper[4894]: I1208 14:47:05.955975 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:05Z","lastTransitionTime":"2025-12-08T14:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.058592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.058625 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.058633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.058646 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.058655 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.161608 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.161654 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.161666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.161684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.161696 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.196525 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.196564 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:06 crc kubenswrapper[4894]: E1208 14:47:06.196657 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:06 crc kubenswrapper[4894]: E1208 14:47:06.196851 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.264038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.264095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.264109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.264128 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.264141 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.365938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.365981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.365993 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.366008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.366019 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.468270 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.468307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.468314 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.468327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.468336 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.571208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.571253 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.571261 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.571277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.571290 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.672950 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.672994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.673004 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.673018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.673028 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.775952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.776003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.776012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.776030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.776040 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.879453 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.879531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.879546 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.879570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.879591 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.983408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.983487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.983505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.983535 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:06 crc kubenswrapper[4894]: I1208 14:47:06.983557 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:06Z","lastTransitionTime":"2025-12-08T14:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.086655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.086694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.086702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.086717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.086728 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.188729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.188807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.188854 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.188880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.188895 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.196336 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:07 crc kubenswrapper[4894]: E1208 14:47:07.196492 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.196577 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:07 crc kubenswrapper[4894]: E1208 14:47:07.196947 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.291323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.291380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.291396 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.291418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.291434 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.393648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.393692 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.393754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.393772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.393784 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.497221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.497265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.497274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.497290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.497303 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.599872 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.599916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.599927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.599945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.599956 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.702972 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.703018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.703035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.703056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.703075 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.805628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.805673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.805683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.805698 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.805711 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.908341 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.908383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.908393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.908410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:07 crc kubenswrapper[4894]: I1208 14:47:07.908423 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:07Z","lastTransitionTime":"2025-12-08T14:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.010757 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.010801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.010810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.010838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.010847 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.113244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.113292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.113300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.113312 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.113321 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.184468 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:08 crc kubenswrapper[4894]: E1208 14:47:08.184619 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:08 crc kubenswrapper[4894]: E1208 14:47:08.184681 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:47:24.184663349 +0000 UTC m=+65.284669484 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.196649 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.196649 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:08 crc kubenswrapper[4894]: E1208 14:47:08.196835 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:08 crc kubenswrapper[4894]: E1208 14:47:08.196895 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.216066 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.216116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.216129 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.216148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.216163 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.267530 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.268528 4894 scope.go:117] "RemoveContainer" containerID="26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be" Dec 08 14:47:08 crc kubenswrapper[4894]: E1208 14:47:08.268710 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.280950 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.291739 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.302046 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.311860 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.318549 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.318593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.318605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.318622 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.318633 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.323464 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.339955 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.353577 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.363916 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.375144 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.384766 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.394702 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.405999 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.416389 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.420388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.420419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.420429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.420442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.420452 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.449029 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.467666 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.485447 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.497213 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:08Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.523048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.523120 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.523130 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.523145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.523154 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.624957 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.624991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.625000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.625013 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.625021 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.727554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.727740 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.727752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.727769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.727781 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.829794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.829858 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.830021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.830039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.830048 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.933166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.933199 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.933207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.933221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:08 crc kubenswrapper[4894]: I1208 14:47:08.933230 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:08Z","lastTransitionTime":"2025-12-08T14:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.035860 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.035903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.035912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.035927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.035937 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.093323 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.093439 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.093472 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:47:41.093454145 +0000 UTC m=+82.193460260 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.093520 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.093536 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.093569 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:41.093561948 +0000 UTC m=+82.193568063 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.093579 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.093625 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:41.093596049 +0000 UTC m=+82.193602164 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.138267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.138306 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.138319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.138334 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.138346 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.194417 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.194487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194599 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194609 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194663 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194678 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194617 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194736 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:41.194717508 +0000 UTC m=+82.294723673 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194740 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.194786 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:47:41.194774319 +0000 UTC m=+82.294780434 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.196387 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.196424 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.196480 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:09 crc kubenswrapper[4894]: E1208 14:47:09.196543 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.207836 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.217743 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.227558 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.235941 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.240922 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.240954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.240964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.240978 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.240988 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.246666 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.257312 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.272673 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.291067 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.304796 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.319734 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.329891 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.340790 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.343421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.343447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.343455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.343469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.343479 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.357421 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.369691 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.382831 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.394111 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.403520 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:09Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.445437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.445469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.445478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.445492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.445499 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.547723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.547797 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.547830 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.547852 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.547865 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.649883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.649926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.649938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.649955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.649965 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.752414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.752467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.752485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.752502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.752569 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.855433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.855477 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.855489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.855510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.855523 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.958407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.958452 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.958463 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.958482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:09 crc kubenswrapper[4894]: I1208 14:47:09.958494 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:09Z","lastTransitionTime":"2025-12-08T14:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.061012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.061053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.061065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.061082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.061093 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.163450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.163488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.163500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.163516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.163526 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.196335 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.196358 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:10 crc kubenswrapper[4894]: E1208 14:47:10.196554 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:10 crc kubenswrapper[4894]: E1208 14:47:10.196790 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.265532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.265574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.265585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.265600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.265610 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.368491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.368529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.368539 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.368556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.368564 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.470247 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.470286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.470303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.470320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.470331 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.573871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.574616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.574680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.574712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.574741 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.677176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.677212 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.677223 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.677238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.677248 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.779765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.779806 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.779833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.779849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.779860 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.882558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.882595 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.882606 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.882621 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.882632 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.985065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.985102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.985109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.985123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:10 crc kubenswrapper[4894]: I1208 14:47:10.985131 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:10Z","lastTransitionTime":"2025-12-08T14:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.087322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.087356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.087365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.087378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.087389 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.189371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.189436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.189448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.189480 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.189498 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.195888 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.195893 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.196094 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.196001 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.228230 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.228269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.228278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.228299 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.228308 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.244026 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:11Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.248573 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.248609 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.248620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.248636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.248648 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.259263 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:11Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.262185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.262207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.262215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.262228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.262237 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.273547 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:11Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.277222 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.277258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.277268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.277283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.277294 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.287263 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:11Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.292326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.292379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.292393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.292411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.292427 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.304544 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:11Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:11 crc kubenswrapper[4894]: E1208 14:47:11.304662 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.306241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.306277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.306287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.306300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.306308 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.409081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.409117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.409125 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.409139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.409148 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.511520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.511558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.511568 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.511582 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.511591 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.613444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.613484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.613495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.613512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.613526 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.716259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.716297 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.716305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.716320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.716329 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.818194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.818226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.818234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.818249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.818257 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.920968 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.921010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.921021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.921038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:11 crc kubenswrapper[4894]: I1208 14:47:11.921049 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:11Z","lastTransitionTime":"2025-12-08T14:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.023755 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.023794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.023802 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.023834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.023847 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.125321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.125359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.125371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.125386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.125395 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.196348 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:12 crc kubenswrapper[4894]: E1208 14:47:12.196503 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.196667 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:12 crc kubenswrapper[4894]: E1208 14:47:12.196720 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.227343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.227382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.227396 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.227416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.227430 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.328377 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.330418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.330485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.330505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.330534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.330553 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.341270 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.342902 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.360166 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.383703 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.397938 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.419671 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.432848 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.433138 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.433207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.433295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.433377 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.435173 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.446129 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.457747 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.470310 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.488193 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.505800 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.519162 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.529438 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.536245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.536310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.536325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.536342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.536353 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.538905 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.547896 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.557849 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.567276 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:12Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.638380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.638415 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.638425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.638440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.638450 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.741018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.741062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.741071 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.741090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.741102 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.842940 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.842980 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.842988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.843003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.843012 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.945291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.945333 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.945343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.945361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:12 crc kubenswrapper[4894]: I1208 14:47:12.945371 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:12Z","lastTransitionTime":"2025-12-08T14:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.047740 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.047780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.047791 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.047807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.047840 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.150061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.150096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.150105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.150120 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.150129 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.196010 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.196013 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:13 crc kubenswrapper[4894]: E1208 14:47:13.196249 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:13 crc kubenswrapper[4894]: E1208 14:47:13.196342 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.252793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.252876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.252894 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.252918 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.252939 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.356123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.356193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.356212 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.356238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.356257 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.459521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.459586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.459625 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.459660 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.459682 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.561319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.561378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.561389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.561404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.561417 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.663958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.663991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.663999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.664012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.664021 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.766861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.766933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.766945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.766964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.766974 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.869773 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.869836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.869848 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.869864 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.869873 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.972765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.972849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.972861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.972880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:13 crc kubenswrapper[4894]: I1208 14:47:13.972895 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:13Z","lastTransitionTime":"2025-12-08T14:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.075437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.076508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.076711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.077041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.077221 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.180584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.180659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.180690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.180720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.180741 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.195881 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.195958 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:14 crc kubenswrapper[4894]: E1208 14:47:14.195994 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:14 crc kubenswrapper[4894]: E1208 14:47:14.196125 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.283215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.283262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.283273 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.283289 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.283299 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.385024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.385064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.385075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.385095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.385106 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.488656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.488724 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.488743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.488765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.488788 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.591411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.591455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.591467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.591483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.591492 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.693750 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.693788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.693799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.693834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.693845 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.796560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.796612 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.796629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.796653 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.796670 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.898751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.898799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.898808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.898842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:14 crc kubenswrapper[4894]: I1208 14:47:14.898852 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:14Z","lastTransitionTime":"2025-12-08T14:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.001522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.001593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.001605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.001642 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.001652 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.103912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.103989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.103998 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.104013 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.104023 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.196653 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.196653 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:15 crc kubenswrapper[4894]: E1208 14:47:15.196793 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:15 crc kubenswrapper[4894]: E1208 14:47:15.196877 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.206372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.206409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.206419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.206431 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.206441 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.308509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.308554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.308565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.308580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.308590 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.411348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.411402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.411419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.411440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.411460 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.514635 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.514707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.514719 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.514737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.514747 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.617845 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.617879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.617887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.617901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.617911 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.720895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.720928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.720937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.720950 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.720959 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.823296 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.823325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.823332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.823347 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.823356 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.925361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.925421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.925433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.925450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:15 crc kubenswrapper[4894]: I1208 14:47:15.925483 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:15Z","lastTransitionTime":"2025-12-08T14:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.027262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.027300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.027310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.027325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.027338 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.130207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.130268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.130280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.130299 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.130913 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.196252 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.196330 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:16 crc kubenswrapper[4894]: E1208 14:47:16.196381 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:16 crc kubenswrapper[4894]: E1208 14:47:16.196520 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.233569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.233605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.233613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.233627 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.233637 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.336132 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.336592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.336712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.337061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.337284 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.440121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.440156 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.440168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.440184 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.440194 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.542938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.542979 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.542990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.543008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.543020 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.645531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.645575 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.645584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.645598 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.645607 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.747900 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.747956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.747967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.747983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.747993 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.850626 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.850659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.850667 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.850682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.850692 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.953494 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.953547 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.953559 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.953578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:16 crc kubenswrapper[4894]: I1208 14:47:16.953590 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:16Z","lastTransitionTime":"2025-12-08T14:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.055686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.055730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.055742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.055758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.055771 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.158050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.158093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.158103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.158118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.158126 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.195973 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.196038 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:17 crc kubenswrapper[4894]: E1208 14:47:17.196103 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:17 crc kubenswrapper[4894]: E1208 14:47:17.196212 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.260489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.260527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.260538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.260556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.260567 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.362896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.362940 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.362949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.362963 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.362972 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.465450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.465495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.465509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.465525 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.465538 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.567352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.567383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.567392 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.567405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.567417 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.670686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.670711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.670720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.670734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.670743 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.773373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.773433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.773444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.773462 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.773472 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.875683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.875747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.875759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.875776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.875788 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.977697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.977742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.977752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.977769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:17 crc kubenswrapper[4894]: I1208 14:47:17.977781 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:17Z","lastTransitionTime":"2025-12-08T14:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.080483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.080520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.080529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.080548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.080558 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.182326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.182381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.182389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.182405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.182414 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.195987 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:18 crc kubenswrapper[4894]: E1208 14:47:18.196153 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.196171 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:18 crc kubenswrapper[4894]: E1208 14:47:18.196291 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.288144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.288192 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.288202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.288221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.288232 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.390541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.390581 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.390588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.390603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.390613 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.493214 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.493265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.493280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.493298 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.493309 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.596383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.596432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.596443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.596460 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.596470 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.700011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.700050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.700067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.700081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.700092 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.802319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.802387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.802405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.802424 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.802433 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.904635 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.904681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.904692 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.904707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:18 crc kubenswrapper[4894]: I1208 14:47:18.904720 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:18Z","lastTransitionTime":"2025-12-08T14:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.007788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.007849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.007862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.007878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.007890 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.110508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.110558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.110574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.110592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.110606 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.196379 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.196596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:19 crc kubenswrapper[4894]: E1208 14:47:19.196914 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:19 crc kubenswrapper[4894]: E1208 14:47:19.196990 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.211938 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.214562 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.214623 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.214635 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.214652 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.214665 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.226347 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.246876 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.261466 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.274315 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.287055 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.298388 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.311809 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.316717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.316758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.316767 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.316800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.316845 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.326750 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.340022 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.350161 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.361039 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.374351 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.386541 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.404555 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.417527 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.420029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.420057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.420065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.420078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.420087 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.432550 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.443005 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:19Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.523336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.523366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.523373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.523387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.523397 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.625696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.625758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.625774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.625799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.625848 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.728259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.728321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.728339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.728398 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.728412 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.833503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.833551 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.833561 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.833586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.833596 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.936893 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.937079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.937094 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.937112 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:19 crc kubenswrapper[4894]: I1208 14:47:19.937125 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:19Z","lastTransitionTime":"2025-12-08T14:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.040581 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.040645 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.040656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.040679 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.040692 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.143491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.143530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.143540 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.143556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.143568 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.195914 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.195936 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:20 crc kubenswrapper[4894]: E1208 14:47:20.196208 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:20 crc kubenswrapper[4894]: E1208 14:47:20.196345 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.246785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.246845 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.246862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.246880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.246890 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.350123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.350173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.350208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.350224 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.350237 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.452427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.452473 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.452486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.452499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.452508 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.555880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.555927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.555936 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.555953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.555962 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.658039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.658087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.658097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.658115 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.658125 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.761749 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.761794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.761803 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.761845 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.761867 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.864212 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.864265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.864274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.864289 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.864299 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.966875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.966912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.966924 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.966939 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:20 crc kubenswrapper[4894]: I1208 14:47:20.966951 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:20Z","lastTransitionTime":"2025-12-08T14:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.069188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.069237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.069250 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.069267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.069332 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.171636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.171674 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.171684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.171698 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.171708 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.196209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.196327 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.196213 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.196615 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.274679 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.274720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.274728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.274744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.274754 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.377379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.377418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.377426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.377441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.377451 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.480035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.480086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.480097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.480114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.480124 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.527712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.527759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.527768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.527785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.527799 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.539565 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:21Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.542721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.542758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.542771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.542786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.542795 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.553510 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:21Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.556782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.556838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.556848 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.556863 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.556873 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.571314 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:21Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.574651 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.574686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.574696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.574712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.574722 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.590809 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:21Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.596141 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.596185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.596198 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.596217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.596230 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.616397 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:21Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:21 crc kubenswrapper[4894]: E1208 14:47:21.616628 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.618318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.618356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.618368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.618385 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.618400 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.721620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.721657 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.721666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.721682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.721692 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.824203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.824228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.824236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.824248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.824257 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.927080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.927116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.927139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.927159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:21 crc kubenswrapper[4894]: I1208 14:47:21.927174 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:21Z","lastTransitionTime":"2025-12-08T14:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.029469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.029498 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.029516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.029532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.029541 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.131804 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.131857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.131867 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.131883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.131894 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.195762 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:22 crc kubenswrapper[4894]: E1208 14:47:22.195922 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.196079 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:22 crc kubenswrapper[4894]: E1208 14:47:22.196133 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.233747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.233773 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.233781 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.233795 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.233804 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.336195 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.336257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.336268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.336283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.336330 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.438961 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.438988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.438996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.439008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.439017 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.541532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.541570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.541578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.541592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.541601 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.644204 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.644239 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.644248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.644263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.644272 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.746574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.746623 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.746636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.746655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.746668 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.849102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.849173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.849187 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.849203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.849214 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.951317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.951364 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.951376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.951392 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:22 crc kubenswrapper[4894]: I1208 14:47:22.951403 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:22Z","lastTransitionTime":"2025-12-08T14:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.053562 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.053607 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.053620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.053639 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.053649 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.156234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.156270 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.156279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.156292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.156301 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.195985 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.195985 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:23 crc kubenswrapper[4894]: E1208 14:47:23.196150 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:23 crc kubenswrapper[4894]: E1208 14:47:23.196216 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.197044 4894 scope.go:117] "RemoveContainer" containerID="26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be" Dec 08 14:47:23 crc kubenswrapper[4894]: E1208 14:47:23.197251 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.258903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.258969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.258984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.259012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.259028 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.361789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.361871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.361884 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.361906 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.361920 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.464246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.464284 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.464292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.464307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.464316 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.566666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.566708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.566718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.566734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.566744 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.669613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.669648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.669658 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.669675 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.669685 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.772123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.772171 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.772179 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.772194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.772203 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.877914 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.877951 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.877961 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.877975 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.877986 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.980616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.980661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.980671 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.980688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:23 crc kubenswrapper[4894]: I1208 14:47:23.980698 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:23Z","lastTransitionTime":"2025-12-08T14:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.083873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.083917 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.083927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.083949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.083962 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.186368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.186420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.186433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.186450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.186462 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.195757 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.195762 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:24 crc kubenswrapper[4894]: E1208 14:47:24.196055 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:24 crc kubenswrapper[4894]: E1208 14:47:24.195920 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.238574 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:24 crc kubenswrapper[4894]: E1208 14:47:24.238720 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:24 crc kubenswrapper[4894]: E1208 14:47:24.238787 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:47:56.238768131 +0000 UTC m=+97.338774246 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.288669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.288712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.288720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.288735 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.288745 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.391829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.391871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.391880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.391896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.391908 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.495759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.495800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.495845 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.495863 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.495874 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.600216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.600248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.600257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.600285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.600295 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.702191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.702235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.702248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.702265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.702277 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.804974 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.805017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.805026 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.805040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.805052 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.908335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.908387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.908397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.908410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:24 crc kubenswrapper[4894]: I1208 14:47:24.908419 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:24Z","lastTransitionTime":"2025-12-08T14:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.009868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.009926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.009938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.009977 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.009991 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.112766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.112965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.113029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.113117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.113176 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.196372 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.196425 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:25 crc kubenswrapper[4894]: E1208 14:47:25.196533 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:25 crc kubenswrapper[4894]: E1208 14:47:25.196714 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.215974 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.216021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.216057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.216076 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.216086 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.318484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.318536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.318546 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.318563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.318581 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.421445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.421485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.421497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.421512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.421520 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.524456 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.524493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.524505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.524522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.524536 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.590701 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/0.log" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.590805 4894 generic.go:334] "Generic (PLEG): container finished" podID="be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235" containerID="6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78" exitCode=1 Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.590885 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerDied","Data":"6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.591435 4894 scope.go:117] "RemoveContainer" containerID="6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.608037 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.625456 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.626991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.627031 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.627041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.627057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.627067 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.640416 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.655109 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.669552 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.679375 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.702992 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.715681 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.730060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.730133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.730148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.730173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.730186 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.736282 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.748516 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.760609 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"2025-12-08T14:46:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b\\\\n2025-12-08T14:46:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b to /host/opt/cni/bin/\\\\n2025-12-08T14:46:40Z [verbose] multus-daemon started\\\\n2025-12-08T14:46:40Z [verbose] Readiness Indicator file check\\\\n2025-12-08T14:47:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.777494 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.790335 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.802907 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.812871 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.822165 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.833073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.833145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.833158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.833176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.833189 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.833715 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.845493 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:25Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.935230 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.935276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.935285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.935299 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:25 crc kubenswrapper[4894]: I1208 14:47:25.935308 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:25Z","lastTransitionTime":"2025-12-08T14:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.037510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.037571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.037585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.037602 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.037613 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.140040 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.140080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.140090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.140107 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.140119 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.196000 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.196058 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:26 crc kubenswrapper[4894]: E1208 14:47:26.196141 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:26 crc kubenswrapper[4894]: E1208 14:47:26.196207 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.241665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.241699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.241709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.241723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.241731 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.344163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.344217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.344228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.344244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.344257 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.446269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.446309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.446323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.446338 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.446349 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.548622 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.548666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.548676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.548691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.548703 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.595542 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/0.log" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.595592 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerStarted","Data":"8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.612622 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.624733 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.633842 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.643484 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.651332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.651370 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.651383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.651400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.651414 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.654704 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.666808 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.677874 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.689456 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.706231 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.719042 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.732279 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.742257 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.753425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.753471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.753483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.753503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.753515 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.756549 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.768195 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.781197 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.792058 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.805016 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"2025-12-08T14:46:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b\\\\n2025-12-08T14:46:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b to /host/opt/cni/bin/\\\\n2025-12-08T14:46:40Z [verbose] multus-daemon started\\\\n2025-12-08T14:46:40Z [verbose] Readiness Indicator file check\\\\n2025-12-08T14:47:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.822569 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:26Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.856047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.856089 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.856100 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.856117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.856127 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.958757 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.958830 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.958842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.958859 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:26 crc kubenswrapper[4894]: I1208 14:47:26.958870 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:26Z","lastTransitionTime":"2025-12-08T14:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.061432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.061471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.061481 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.061497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.061507 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.163965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.164019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.164028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.164048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.164062 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.196573 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.196639 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:27 crc kubenswrapper[4894]: E1208 14:47:27.196687 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:27 crc kubenswrapper[4894]: E1208 14:47:27.196764 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.267113 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.267157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.267168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.267188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.267209 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.369715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.369761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.369774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.369794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.369805 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.471777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.471829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.471838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.471853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.471862 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.573764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.573804 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.573832 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.573849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.573861 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.676008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.676055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.676068 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.676086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.676098 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.778654 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.778693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.778707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.778724 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.778736 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.881340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.881375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.881384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.881397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.881406 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.983734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.983796 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.983807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.983868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:27 crc kubenswrapper[4894]: I1208 14:47:27.983880 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:27Z","lastTransitionTime":"2025-12-08T14:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.086061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.086098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.086106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.086121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.086132 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.188795 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.188853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.188864 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.188878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.188888 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.196428 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.196453 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:28 crc kubenswrapper[4894]: E1208 14:47:28.196573 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:28 crc kubenswrapper[4894]: E1208 14:47:28.196669 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.292280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.292328 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.292341 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.292359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.292375 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.395177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.395219 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.395228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.395249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.395267 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.497172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.497216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.497229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.497245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.497258 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.600353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.600393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.600402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.600416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.600427 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.702874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.702915 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.702927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.702946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.702961 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.805418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.805467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.805482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.805501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.805512 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.908301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.908338 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.908349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.908376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:28 crc kubenswrapper[4894]: I1208 14:47:28.908388 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:28Z","lastTransitionTime":"2025-12-08T14:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.010457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.010501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.010512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.010530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.010541 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.112706 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.112760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.112786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.112808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.112836 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.196085 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.196155 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:29 crc kubenswrapper[4894]: E1208 14:47:29.196271 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:29 crc kubenswrapper[4894]: E1208 14:47:29.196367 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.208668 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.214600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.214628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.214638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.214657 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.214668 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.219223 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.229043 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.239108 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.253369 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.265029 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.276140 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.286284 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.299083 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.313192 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.316520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.316558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.316572 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.316597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.316611 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.325964 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.342846 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.354682 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.364930 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.373853 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.386725 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"2025-12-08T14:46:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b\\\\n2025-12-08T14:46:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b to /host/opt/cni/bin/\\\\n2025-12-08T14:46:40Z [verbose] multus-daemon started\\\\n2025-12-08T14:46:40Z [verbose] Readiness Indicator file check\\\\n2025-12-08T14:47:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.403378 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.416954 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:29Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.418522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.418566 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.418576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.418594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.418604 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.520638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.520680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.520693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.520708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.520720 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.623199 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.623246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.623286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.623307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.623320 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.725913 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.725953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.725963 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.725978 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.725987 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.828106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.828143 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.828155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.828173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.828184 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.930682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.930729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.930740 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.930761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:29 crc kubenswrapper[4894]: I1208 14:47:29.930772 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:29Z","lastTransitionTime":"2025-12-08T14:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.033211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.033257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.033290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.033308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.033320 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.136676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.136786 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.136857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.136874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.136899 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.196502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.196548 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:30 crc kubenswrapper[4894]: E1208 14:47:30.196648 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:30 crc kubenswrapper[4894]: E1208 14:47:30.196916 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.209896 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.239461 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.239505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.239517 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.239535 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.239546 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.342953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.343004 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.343019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.343038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.343052 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.445330 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.445361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.445369 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.445382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.445392 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.547476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.547520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.547531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.547547 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.547560 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.650229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.650283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.650297 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.650315 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.650326 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.753342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.753425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.753435 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.753473 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.753486 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.856323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.856585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.856733 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.856869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.856989 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.959987 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.960052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.960065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.960089 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:30 crc kubenswrapper[4894]: I1208 14:47:30.960101 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:30Z","lastTransitionTime":"2025-12-08T14:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.062277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.062587 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.062711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.062791 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.062885 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.165888 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.166218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.166386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.166474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.166567 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.196447 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.196520 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.196826 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.197076 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.268715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.268761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.268771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.268785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.268794 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.371496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.371534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.371546 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.371562 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.371574 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.473908 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.473944 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.473954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.473970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.473980 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.576078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.576117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.576127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.576141 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.576151 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.678680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.678722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.678731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.678744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.678754 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.705055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.705084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.705093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.705105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.705114 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.715829 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:31Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.719603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.719631 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.719640 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.719656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.719666 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.731547 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:31Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.737101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.737139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.737152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.737172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.737182 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.756580 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:31Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.760469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.760500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.760508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.760522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.760534 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.772990 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:31Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.776054 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.776093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.776102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.776116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.776126 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.791308 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"058a8d66-2b8f-482d-b95a-6b5107a7954b\\\",\\\"systemUUID\\\":\\\"0f545b9e-425e-49eb-9611-92fe15e98ca0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:31Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:31 crc kubenswrapper[4894]: E1208 14:47:31.791455 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.793124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.793176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.793189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.793205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.793235 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.895209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.895272 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.895304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.895321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.895352 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.997548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.997904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.998024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.998113 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:31 crc kubenswrapper[4894]: I1208 14:47:31.998179 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:31Z","lastTransitionTime":"2025-12-08T14:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.100928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.101158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.101259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.101333 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.101403 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.196129 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.196207 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:32 crc kubenswrapper[4894]: E1208 14:47:32.196672 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:32 crc kubenswrapper[4894]: E1208 14:47:32.196520 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.204078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.204111 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.204123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.204137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.204148 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.306301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.306339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.306350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.306365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.306375 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.409122 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.409144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.409152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.409165 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.409174 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.511465 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.511500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.511510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.511527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.511538 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.614555 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.615158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.615242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.615305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.615369 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.718177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.718211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.718219 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.718250 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.718259 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.820277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.820319 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.820332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.820348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.820359 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.922512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.922540 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.922547 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.922562 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:32 crc kubenswrapper[4894]: I1208 14:47:32.922570 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:32Z","lastTransitionTime":"2025-12-08T14:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.024542 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.024572 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.024580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.024594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.024602 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.126345 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.126388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.126400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.126416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.126426 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.195901 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:33 crc kubenswrapper[4894]: E1208 14:47:33.196244 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.196284 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:33 crc kubenswrapper[4894]: E1208 14:47:33.196434 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.228545 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.228597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.228608 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.228629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.228640 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.330309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.330348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.330361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.330377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.330388 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.433276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.433306 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.433337 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.433351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.433362 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.536130 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.536163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.536174 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.536190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.536200 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.638996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.639037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.639045 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.639059 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.639069 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.742613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.742654 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.742663 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.742678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.742688 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.844923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.844962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.844977 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.844992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.845003 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.946927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.946958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.946969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.947014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:33 crc kubenswrapper[4894]: I1208 14:47:33.947032 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:33Z","lastTransitionTime":"2025-12-08T14:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.049598 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.049636 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.049648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.049665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.049676 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.152571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.152616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.152630 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.152648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.152660 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.196305 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.196397 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:34 crc kubenswrapper[4894]: E1208 14:47:34.196490 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:34 crc kubenswrapper[4894]: E1208 14:47:34.196572 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.254876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.254944 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.254967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.254996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.255018 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.357694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.357733 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.357741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.357755 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.357764 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.460134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.460230 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.460248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.460266 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.460283 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.563163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.563200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.563211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.563229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.563242 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.665224 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.665269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.665279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.665294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.665303 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.767675 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.767722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.767731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.767746 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.767756 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.869869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.869901 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.869910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.869922 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.869931 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.972443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.972476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.972484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.972498 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:34 crc kubenswrapper[4894]: I1208 14:47:34.972506 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:34Z","lastTransitionTime":"2025-12-08T14:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.074709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.074745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.074756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.074773 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.074785 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.177282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.177331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.177347 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.177362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.177379 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.195893 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.195922 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:35 crc kubenswrapper[4894]: E1208 14:47:35.196031 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:35 crc kubenswrapper[4894]: E1208 14:47:35.196102 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.279756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.279803 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.279833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.279849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.279861 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.382392 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.382438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.382450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.382468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.382480 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.485059 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.485107 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.485120 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.485138 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.485151 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.588075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.588112 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.588124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.588142 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.588155 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.690558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.690607 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.690615 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.690628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.690637 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.792711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.792749 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.792758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.792774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.792793 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.894754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.894805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.894853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.894873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.894892 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.997273 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.997333 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.997349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.997368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:35 crc kubenswrapper[4894]: I1208 14:47:35.997380 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:35Z","lastTransitionTime":"2025-12-08T14:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.100181 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.100234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.100250 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.100269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.100284 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.196356 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.196417 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:36 crc kubenswrapper[4894]: E1208 14:47:36.196503 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:36 crc kubenswrapper[4894]: E1208 14:47:36.196566 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.202373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.202416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.202427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.202443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.202452 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.304664 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.304711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.304723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.304742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.304752 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.407381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.407419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.407430 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.407447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.407458 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.509760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.509798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.509808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.509836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.509844 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.611666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.611707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.611717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.611735 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.611745 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.714868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.714932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.714946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.714964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.714978 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.816984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.817022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.817029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.817043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.817051 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.919256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.919295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.919307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.919323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:36 crc kubenswrapper[4894]: I1208 14:47:36.919334 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:36Z","lastTransitionTime":"2025-12-08T14:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.022321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.022388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.022404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.022425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.022441 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.125432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.125478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.125490 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.125506 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.125517 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.196646 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.196728 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:37 crc kubenswrapper[4894]: E1208 14:47:37.196777 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:37 crc kubenswrapper[4894]: E1208 14:47:37.196981 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.197764 4894 scope.go:117] "RemoveContainer" containerID="26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.228797 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.228842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.228852 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.228866 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.228874 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.331042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.331079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.331090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.331105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.331117 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.433237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.433288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.433308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.433321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.433330 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.535809 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.535873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.535883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.535900 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.535912 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.637965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.638007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.638017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.638031 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.638039 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.740491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.740563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.740571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.740591 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.740600 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.843235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.843278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.843291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.843308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.843317 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.945751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.945793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.945802 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.945832 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:37 crc kubenswrapper[4894]: I1208 14:47:37.945843 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:37Z","lastTransitionTime":"2025-12-08T14:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.048529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.048564 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.048576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.048591 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.048602 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.183833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.183868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.183876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.183892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.183901 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.196108 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.196183 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:38 crc kubenswrapper[4894]: E1208 14:47:38.196247 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:38 crc kubenswrapper[4894]: E1208 14:47:38.196315 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.285615 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.285653 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.285661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.285677 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.285687 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.387767 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.387834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.387844 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.387865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.387887 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.489910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.489962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.489980 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.489997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.490008 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.592160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.592198 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.592207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.592222 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.592231 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.635347 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/2.log" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.638159 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.638557 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.657549 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.668707 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.683052 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.694341 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.694386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.694398 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.694414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.694427 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.694809 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.706623 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"2025-12-08T14:46:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b\\\\n2025-12-08T14:46:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b to /host/opt/cni/bin/\\\\n2025-12-08T14:46:40Z [verbose] multus-daemon started\\\\n2025-12-08T14:46:40Z [verbose] Readiness Indicator file check\\\\n2025-12-08T14:47:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.724591 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.737625 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.748728 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.759365 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"146f7e76-6335-4d7b-bc5a-e7f4ce63175f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ebdc46c969cd502cf4084c56a87b1129f69d0c8b3da96efc447f6be04be840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.770146 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.779314 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.792075 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.797737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.797779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.797789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.797803 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.797831 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.803622 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.816095 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.825619 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.836949 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.847620 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.860032 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.872587 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.900480 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.900520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.900529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.900543 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:38 crc kubenswrapper[4894]: I1208 14:47:38.900553 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:38Z","lastTransitionTime":"2025-12-08T14:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.002387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.002444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.002452 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.002464 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.002472 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.104774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.104799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.104807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.104850 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.104864 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.198106 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.198149 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:39 crc kubenswrapper[4894]: E1208 14:47:39.198344 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:39 crc kubenswrapper[4894]: E1208 14:47:39.198455 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.207188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.207224 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.207232 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.207245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.207254 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.213251 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.224280 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.233977 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.243239 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.250690 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.260695 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.273681 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.287387 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.309186 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.309474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.309498 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.309509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.309523 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.309534 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.323300 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.338591 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.348758 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.358110 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.371905 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"2025-12-08T14:46:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b\\\\n2025-12-08T14:46:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b to /host/opt/cni/bin/\\\\n2025-12-08T14:46:40Z [verbose] multus-daemon started\\\\n2025-12-08T14:46:40Z [verbose] Readiness Indicator file check\\\\n2025-12-08T14:47:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.389970 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.402013 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.411779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.411834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.411846 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.411863 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.411876 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.413999 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.424121 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"146f7e76-6335-4d7b-bc5a-e7f4ce63175f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ebdc46c969cd502cf4084c56a87b1129f69d0c8b3da96efc447f6be04be840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.434101 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.514929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.514971 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.514982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.515000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.515012 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.616737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.616767 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.616775 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.616789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.616798 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.642862 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/3.log" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.643483 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/2.log" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.645983 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189" exitCode=1 Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.646037 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.646103 4894 scope.go:117] "RemoveContainer" containerID="26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.646879 4894 scope.go:117] "RemoveContainer" containerID="d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189" Dec 08 14:47:39 crc kubenswrapper[4894]: E1208 14:47:39.647038 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.667424 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.680114 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.702285 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.714473 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.718949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.718995 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.719012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.719029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.719042 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.727043 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.740208 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.752781 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"146f7e76-6335-4d7b-bc5a-e7f4ce63175f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ebdc46c969cd502cf4084c56a87b1129f69d0c8b3da96efc447f6be04be840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.772461 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.785417 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.797841 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"2025-12-08T14:46:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b\\\\n2025-12-08T14:46:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b to /host/opt/cni/bin/\\\\n2025-12-08T14:46:40Z [verbose] multus-daemon started\\\\n2025-12-08T14:46:40Z [verbose] Readiness Indicator file check\\\\n2025-12-08T14:47:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.815173 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://26823ebf14b72e83f263ba932cfb15a55bfd2c6e8f43366db351dba99ba029be\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:03Z\\\",\\\"message\\\":\\\"03] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nF1208 14:47:03.050051 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:02Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:03.050364 6512 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1208 14:47:03.050371 6512 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1208 14:47:03.050356 6512 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSe\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:38Z\\\",\\\"message\\\":\\\"or/machine-config-daemon-97dqr\\\\nI1208 14:47:38.562402 6882 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nF1208 14:47:38.562404 6882 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:38.562408 6882 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-97dqr in node crc\\\\nI1208 14:47:38.562417 6882 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-97dqr after 0 failed attempt(s)\\\\nI1208 14:47:38.562419 6882 obj_retry.go:303]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.821737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.821984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.822051 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.822119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.822185 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.830139 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.843320 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.856181 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.866029 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.876128 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.886605 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.897861 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.911169 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:39Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.924698 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.924747 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.924756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.924769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:39 crc kubenswrapper[4894]: I1208 14:47:39.924778 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:39Z","lastTransitionTime":"2025-12-08T14:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.027436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.027483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.027492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.027507 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.027518 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.130351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.130379 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.130386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.130400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.130408 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.196031 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:40 crc kubenswrapper[4894]: E1208 14:47:40.196158 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.196031 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:40 crc kubenswrapper[4894]: E1208 14:47:40.196346 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.232529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.232905 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.232978 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.233111 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.233184 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.336018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.336052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.336063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.336080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.336092 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.439264 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.439333 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.439349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.439376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.439392 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.542024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.542082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.542096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.542120 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.542133 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.645397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.645741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.645865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.645970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.646054 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.652684 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/3.log" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.657321 4894 scope.go:117] "RemoveContainer" containerID="d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189" Dec 08 14:47:40 crc kubenswrapper[4894]: E1208 14:47:40.657544 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.669418 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5422fa4f-2895-4053-91cb-617b0c903beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9baf30dbc598e5166726ffcc3fa250032fb7308c2aad563743ee70f7f579976a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5998e9516f8831d827e48ada9e70ca8a550b8bc29fbd3f70a46b1e639058bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://214ee123fb832e213956a18e5032bf10e27dd6089b0c6a403edb0b0c4dd13f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b6f60b71e01b4c1d38515dfeb7e47b97f196b6651531cb0f323b569f446d606\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.684801 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.700625 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df379edacf4ea3271faf7eb866e166f9c46e74507ebe1e8140e60df1665b93d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.734477 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34aea25d-04e6-4666-88f8-df004542be30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c43ab538464896ca7b22fb1dd0258bae75eb75b35410721bf824ab254552e75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05461a6de44bee9733de2f843928f1eb41e113d8921125edf3ac5829c044cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b2ea92cbda151939d8c0cb3fd738cda5b6e09cca1dc02b112d5f52ac7225afc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14e2cd9d817596cfb4eb1025106633d42cc07c7df2f43b999aea34f62f56ba5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a81a7708b889947d9a56abfb049d8c415038e030bea951d66de60c842ec06448\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d868b7ee02eea5b4ca87ff3c8732b5b172a7af8e4aa89bedd504e9324ee84f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2f3a7bb6b774ebe6de533d24200c530a6d212e17f80df3734e6792bc5a6d47a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c993769c1783a4ff2fa3609294d0434c3ea305b1bd03514d4887eddd77526a43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.749495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.749799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.749928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.750088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.750201 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.751299 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.802183 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pbxln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2c74a28-06eb-4b6f-b540-d83b4f69f2a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e94da1f80cc821ee0096a3101fa44e5cc1b5c67d5c3a0fb543876bebd632e70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://357921021a48ef09938cba7bd91fb7afd16fdb2f10422846ceddb3f276ce115d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8dffad57362401c1854d881f0f77292a23c68a989bb29f46187714743b8fe996\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://852701f4698270575253563de8ee3d4359dca5e589d489cab0bfed88636d54e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://368395fdd9fcb628bcb4229e2260b624fa936a153ca6d4297b6c45f05661f334\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7336cec03c9b7bf778f9ee68f2f6b5aee9651493e6309186556076421d70a1e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c9dc390cb3e60863d6dff64ae36571acbc5b621ad318573c53b9882c370d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-68w6k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pbxln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.819330 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49a05fc4-1361-4ebc-891a-e2b49df28ffa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvwm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2zz6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.839904 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb212a28-36c5-440e-8965-986352c5d3ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:38Z\\\",\\\"message\\\":\\\"or/machine-config-daemon-97dqr\\\\nI1208 14:47:38.562402 6882 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-scheduler/openshift-kube-scheduler-crc\\\\nF1208 14:47:38.562404 6882 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:38Z is after 2025-08-24T17:21:41Z]\\\\nI1208 14:47:38.562408 6882 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-97dqr in node crc\\\\nI1208 14:47:38.562417 6882 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-97dqr after 0 failed attempt(s)\\\\nI1208 14:47:38.562419 6882 obj_retry.go:303]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:47:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6qpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c25bk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.853343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.853377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.853386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.853407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.853417 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.856489 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9112ccff-5eb4-44a5-b333-dac4d0474d73\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\":21.855637 1 observer_polling.go:159] Starting file observer\\\\nW1208 14:46:31.858294 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1208 14:46:31.858492 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1208 14:46:31.859881 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-951518102/tls.crt::/tmp/serving-cert-951518102/tls.key\\\\\\\"\\\\nI1208 14:46:37.212534 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1208 14:46:37.215023 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1208 14:46:37.215046 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1208 14:46:37.215065 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1208 14:46:37.215072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1208 14:46:37.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1208 14:46:37.228619 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228657 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1208 14:46:37.228689 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1208 14:46:37.228921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1208 14:46:37.228981 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1208 14:46:37.229258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1208 14:46:37.229505 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.876832 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a6d4a91-bf94-4ca7-961e-1266414ce304\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9966dfe57cd540940d6df87889e0d40efcfcf20e9477f3416e9cb0cecf367c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dc6b70702558c20bf3bd809ae0a80e5b3eea3e300cec6fcc35be20ba2c40503\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c98065a0478b27b181acd1385e507e57c1b81f89b9a99dbb008f4c254819344b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.891102 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"146f7e76-6335-4d7b-bc5a-e7f4ce63175f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ebdc46c969cd502cf4084c56a87b1129f69d0c8b3da96efc447f6be04be840\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb4d5b2b1a26101d966b3e89de6a9865531c05e05c0c3c19e0811058eb9799d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-08T14:46:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.903213 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b42bc6b0baafd6bdb5ebfc6011faed588fccca92040a39761cce181e13d6735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.915744 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rr2kq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f4dd416-8f0d-4efb-944d-1720665a8d2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d04db78654b4b090a2dc651849658806b8def4d9e2560198d2b1d0970d7be7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-75zfw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rr2kq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.931912 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jtx8z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:47:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-08T14:47:25Z\\\",\\\"message\\\":\\\"2025-12-08T14:46:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b\\\\n2025-12-08T14:46:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8f51dfa0-b079-44f7-9e34-96d5eef4017b to /host/opt/cni/bin/\\\\n2025-12-08T14:46:40Z [verbose] multus-daemon started\\\\n2025-12-08T14:46:40Z [verbose] Readiness Indicator file check\\\\n2025-12-08T14:47:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:47:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tmnss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jtx8z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.947755 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd66f4628d6362b92b980b1d06266e2c15280f2895d3cff77ffb18f7ca0318e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23b07fa348543754c8e8d140db446264b600178ecbefe0b1428543edc89b11f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.956122 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.956190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.956203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.956223 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.956236 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:40Z","lastTransitionTime":"2025-12-08T14:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.963317 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.977866 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b27019e5-2a3d-414e-b2ee-7606492ba074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://108724a388112661c7400831427e09fbcd05ede3ed05d08f2941e24308dea6fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jqtll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-97dqr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:40 crc kubenswrapper[4894]: I1208 14:47:40.988746 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-z6s8q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60593d51-c757-4138-855d-6904f15385b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cedf3f6ac33422a21f375db8df873c5484ef7aae3ef8048294c64cfb1f9ac4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xnnfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-z6s8q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.002236 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e7a706-dfa9-4c32-a7bd-478d8faa771b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-08T14:46:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2e9cf67d84c378ca0909971de8c38c53f3f8f9b71b619729b5413bb7bee38bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8371d5838d336102cb921dad2d2b0344ad7d2e0c7c370aa8ed4edc06c7dba02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-08T14:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-994vg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-08T14:46:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-zgtl5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-08T14:47:40Z is after 2025-08-24T17:21:41Z" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.064043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.064086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.064095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.064117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.064130 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.110073 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.110213 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.110276 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:45.110252962 +0000 UTC m=+146.210259087 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.110301 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.110347 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:48:45.110336274 +0000 UTC m=+146.210342389 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.110366 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.110474 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.110508 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-08 14:48:45.11049856 +0000 UTC m=+146.210504675 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.166126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.166177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.166190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.166208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.166221 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.196865 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.196998 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.197166 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.197467 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.211249 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.211321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211470 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211499 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211497 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211511 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211522 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211536 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211569 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-08 14:48:45.211551841 +0000 UTC m=+146.311557956 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:41 crc kubenswrapper[4894]: E1208 14:47:41.211596 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-08 14:48:45.211577182 +0000 UTC m=+146.311583297 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.269585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.269945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.269954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.269968 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.269977 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.373372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.373412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.373421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.373438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.373447 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.476119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.476172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.476185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.476200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.476212 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.584308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.584366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.584381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.584403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.584416 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.686958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.687010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.687020 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.687039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.687051 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.790060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.790124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.790137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.790159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.790173 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.827046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.827088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.827098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.827117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.827128 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-08T14:47:41Z","lastTransitionTime":"2025-12-08T14:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.878631 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4"] Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.879032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.880932 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.881350 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.881355 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.883380 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.895452 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=29.895410081 podStartE2EDuration="29.895410081s" podCreationTimestamp="2025-12-08 14:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:41.894972357 +0000 UTC m=+82.994978482" watchObservedRunningTime="2025-12-08 14:47:41.895410081 +0000 UTC m=+82.995416196" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.957487 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=62.95746991 podStartE2EDuration="1m2.95746991s" podCreationTimestamp="2025-12-08 14:46:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:41.956401286 +0000 UTC m=+83.056407401" watchObservedRunningTime="2025-12-08 14:47:41.95746991 +0000 UTC m=+83.057476025" Dec 08 14:47:41 crc kubenswrapper[4894]: I1208 14:47:41.997104 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pbxln" podStartSLOduration=63.997084494 podStartE2EDuration="1m3.997084494s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:41.98598286 +0000 UTC m=+83.085988975" watchObservedRunningTime="2025-12-08 14:47:41.997084494 +0000 UTC m=+83.097090609" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.012338 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=65.012317139 podStartE2EDuration="1m5.012317139s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.011337849 +0000 UTC m=+83.111343964" watchObservedRunningTime="2025-12-08 14:47:42.012317139 +0000 UTC m=+83.112323264" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.018278 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dd609389-73e4-4c8a-9c6e-a1f7575d0550-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.018339 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd609389-73e4-4c8a-9c6e-a1f7575d0550-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.018391 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dd609389-73e4-4c8a-9c6e-a1f7575d0550-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.018412 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dd609389-73e4-4c8a-9c6e-a1f7575d0550-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.018435 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd609389-73e4-4c8a-9c6e-a1f7575d0550-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.035790 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=61.035772478 podStartE2EDuration="1m1.035772478s" podCreationTimestamp="2025-12-08 14:46:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.024983394 +0000 UTC m=+83.124989509" watchObservedRunningTime="2025-12-08 14:47:42.035772478 +0000 UTC m=+83.135778593" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.036026 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=12.036019416 podStartE2EDuration="12.036019416s" podCreationTimestamp="2025-12-08 14:47:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.035979374 +0000 UTC m=+83.135985489" watchObservedRunningTime="2025-12-08 14:47:42.036019416 +0000 UTC m=+83.136025531" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.067894 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rr2kq" podStartSLOduration=65.067878122 podStartE2EDuration="1m5.067878122s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.054555476 +0000 UTC m=+83.154561591" watchObservedRunningTime="2025-12-08 14:47:42.067878122 +0000 UTC m=+83.167884247" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.068202 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jtx8z" podStartSLOduration=64.068198181 podStartE2EDuration="1m4.068198181s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.068097009 +0000 UTC m=+83.168103124" watchObservedRunningTime="2025-12-08 14:47:42.068198181 +0000 UTC m=+83.168204296" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.118900 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd609389-73e4-4c8a-9c6e-a1f7575d0550-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.118958 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dd609389-73e4-4c8a-9c6e-a1f7575d0550-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.118975 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dd609389-73e4-4c8a-9c6e-a1f7575d0550-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.118993 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd609389-73e4-4c8a-9c6e-a1f7575d0550-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.119046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dd609389-73e4-4c8a-9c6e-a1f7575d0550-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.119098 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dd609389-73e4-4c8a-9c6e-a1f7575d0550-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.119137 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dd609389-73e4-4c8a-9c6e-a1f7575d0550-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.120051 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dd609389-73e4-4c8a-9c6e-a1f7575d0550-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.124983 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd609389-73e4-4c8a-9c6e-a1f7575d0550-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.130206 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podStartSLOduration=65.130189518 podStartE2EDuration="1m5.130189518s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.130127286 +0000 UTC m=+83.230133411" watchObservedRunningTime="2025-12-08 14:47:42.130189518 +0000 UTC m=+83.230195633" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.139642 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd609389-73e4-4c8a-9c6e-a1f7575d0550-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bgzt4\" (UID: \"dd609389-73e4-4c8a-9c6e-a1f7575d0550\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.140222 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-z6s8q" podStartSLOduration=65.140202288 podStartE2EDuration="1m5.140202288s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.139368071 +0000 UTC m=+83.239374186" watchObservedRunningTime="2025-12-08 14:47:42.140202288 +0000 UTC m=+83.240208403" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.150534 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-zgtl5" podStartSLOduration=64.150519157 podStartE2EDuration="1m4.150519157s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.149424602 +0000 UTC m=+83.249430717" watchObservedRunningTime="2025-12-08 14:47:42.150519157 +0000 UTC m=+83.250525272" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.194500 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.195799 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.195963 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:42 crc kubenswrapper[4894]: E1208 14:47:42.197017 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:42 crc kubenswrapper[4894]: E1208 14:47:42.197150 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:42 crc kubenswrapper[4894]: W1208 14:47:42.210370 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd609389_73e4_4c8a_9c6e_a1f7575d0550.slice/crio-8febaaaf4796aa21353bf5d73df65709c2caa5859c467fe8e65c0b578e3b0a4b WatchSource:0}: Error finding container 8febaaaf4796aa21353bf5d73df65709c2caa5859c467fe8e65c0b578e3b0a4b: Status 404 returned error can't find the container with id 8febaaaf4796aa21353bf5d73df65709c2caa5859c467fe8e65c0b578e3b0a4b Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.664364 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" event={"ID":"dd609389-73e4-4c8a-9c6e-a1f7575d0550","Type":"ContainerStarted","Data":"6ef98abba6d4bf0b63735a966140d12c7b961e95c2882cb44ab6a7fc5600c4a9"} Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.664422 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" event={"ID":"dd609389-73e4-4c8a-9c6e-a1f7575d0550","Type":"ContainerStarted","Data":"8febaaaf4796aa21353bf5d73df65709c2caa5859c467fe8e65c0b578e3b0a4b"} Dec 08 14:47:42 crc kubenswrapper[4894]: I1208 14:47:42.678303 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bgzt4" podStartSLOduration=64.678279569 podStartE2EDuration="1m4.678279569s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:47:42.677548205 +0000 UTC m=+83.777554330" watchObservedRunningTime="2025-12-08 14:47:42.678279569 +0000 UTC m=+83.778285684" Dec 08 14:47:43 crc kubenswrapper[4894]: I1208 14:47:43.195768 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:43 crc kubenswrapper[4894]: I1208 14:47:43.195887 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:43 crc kubenswrapper[4894]: E1208 14:47:43.196064 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:43 crc kubenswrapper[4894]: E1208 14:47:43.196182 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:44 crc kubenswrapper[4894]: I1208 14:47:44.196171 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:44 crc kubenswrapper[4894]: I1208 14:47:44.196200 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:44 crc kubenswrapper[4894]: E1208 14:47:44.196338 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:44 crc kubenswrapper[4894]: E1208 14:47:44.196441 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:45 crc kubenswrapper[4894]: I1208 14:47:45.196338 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:45 crc kubenswrapper[4894]: I1208 14:47:45.196455 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:45 crc kubenswrapper[4894]: E1208 14:47:45.196759 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:45 crc kubenswrapper[4894]: E1208 14:47:45.197008 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:46 crc kubenswrapper[4894]: I1208 14:47:46.196778 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:46 crc kubenswrapper[4894]: I1208 14:47:46.196840 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:46 crc kubenswrapper[4894]: E1208 14:47:46.197342 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:46 crc kubenswrapper[4894]: E1208 14:47:46.197802 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:47 crc kubenswrapper[4894]: I1208 14:47:47.196771 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:47 crc kubenswrapper[4894]: I1208 14:47:47.196870 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:47 crc kubenswrapper[4894]: E1208 14:47:47.197248 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:47 crc kubenswrapper[4894]: E1208 14:47:47.197305 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:48 crc kubenswrapper[4894]: I1208 14:47:48.195981 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:48 crc kubenswrapper[4894]: I1208 14:47:48.196131 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:48 crc kubenswrapper[4894]: E1208 14:47:48.196170 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:48 crc kubenswrapper[4894]: E1208 14:47:48.196598 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:49 crc kubenswrapper[4894]: I1208 14:47:49.197364 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:49 crc kubenswrapper[4894]: I1208 14:47:49.197427 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:49 crc kubenswrapper[4894]: E1208 14:47:49.197499 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:49 crc kubenswrapper[4894]: E1208 14:47:49.197623 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:50 crc kubenswrapper[4894]: I1208 14:47:50.196720 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:50 crc kubenswrapper[4894]: I1208 14:47:50.196751 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:50 crc kubenswrapper[4894]: E1208 14:47:50.196995 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:50 crc kubenswrapper[4894]: E1208 14:47:50.197256 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:51 crc kubenswrapper[4894]: I1208 14:47:51.196217 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:51 crc kubenswrapper[4894]: I1208 14:47:51.196318 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:51 crc kubenswrapper[4894]: E1208 14:47:51.196360 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:51 crc kubenswrapper[4894]: E1208 14:47:51.196560 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:52 crc kubenswrapper[4894]: I1208 14:47:52.196615 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:52 crc kubenswrapper[4894]: I1208 14:47:52.196856 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:52 crc kubenswrapper[4894]: E1208 14:47:52.196895 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:52 crc kubenswrapper[4894]: E1208 14:47:52.197129 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:53 crc kubenswrapper[4894]: I1208 14:47:53.195906 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:53 crc kubenswrapper[4894]: I1208 14:47:53.196003 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:53 crc kubenswrapper[4894]: E1208 14:47:53.196068 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:53 crc kubenswrapper[4894]: E1208 14:47:53.196150 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:54 crc kubenswrapper[4894]: I1208 14:47:54.196793 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:54 crc kubenswrapper[4894]: E1208 14:47:54.197309 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:54 crc kubenswrapper[4894]: I1208 14:47:54.197322 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:54 crc kubenswrapper[4894]: E1208 14:47:54.197641 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:54 crc kubenswrapper[4894]: I1208 14:47:54.197925 4894 scope.go:117] "RemoveContainer" containerID="d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189" Dec 08 14:47:54 crc kubenswrapper[4894]: E1208 14:47:54.198154 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:47:55 crc kubenswrapper[4894]: I1208 14:47:55.196015 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:55 crc kubenswrapper[4894]: E1208 14:47:55.196138 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:55 crc kubenswrapper[4894]: I1208 14:47:55.196183 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:55 crc kubenswrapper[4894]: E1208 14:47:55.196235 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:56 crc kubenswrapper[4894]: I1208 14:47:56.196445 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:56 crc kubenswrapper[4894]: I1208 14:47:56.196525 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:56 crc kubenswrapper[4894]: E1208 14:47:56.196595 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:56 crc kubenswrapper[4894]: E1208 14:47:56.196686 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:56 crc kubenswrapper[4894]: I1208 14:47:56.280910 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:56 crc kubenswrapper[4894]: E1208 14:47:56.281035 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:56 crc kubenswrapper[4894]: E1208 14:47:56.281113 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs podName:49a05fc4-1361-4ebc-891a-e2b49df28ffa nodeName:}" failed. No retries permitted until 2025-12-08 14:49:00.281092497 +0000 UTC m=+161.381098602 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs") pod "network-metrics-daemon-z2zz6" (UID: "49a05fc4-1361-4ebc-891a-e2b49df28ffa") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 08 14:47:57 crc kubenswrapper[4894]: I1208 14:47:57.196283 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:57 crc kubenswrapper[4894]: I1208 14:47:57.196366 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:57 crc kubenswrapper[4894]: E1208 14:47:57.196430 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:47:57 crc kubenswrapper[4894]: E1208 14:47:57.196602 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:58 crc kubenswrapper[4894]: I1208 14:47:58.196517 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:47:58 crc kubenswrapper[4894]: E1208 14:47:58.196721 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:47:58 crc kubenswrapper[4894]: I1208 14:47:58.196529 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:47:58 crc kubenswrapper[4894]: E1208 14:47:58.197146 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:47:59 crc kubenswrapper[4894]: I1208 14:47:59.196156 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:47:59 crc kubenswrapper[4894]: I1208 14:47:59.196188 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:47:59 crc kubenswrapper[4894]: E1208 14:47:59.197263 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:47:59 crc kubenswrapper[4894]: E1208 14:47:59.197308 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:00 crc kubenswrapper[4894]: I1208 14:48:00.196023 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:00 crc kubenswrapper[4894]: I1208 14:48:00.196030 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:00 crc kubenswrapper[4894]: E1208 14:48:00.196164 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:00 crc kubenswrapper[4894]: E1208 14:48:00.196354 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:01 crc kubenswrapper[4894]: I1208 14:48:01.196637 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:01 crc kubenswrapper[4894]: I1208 14:48:01.196637 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:01 crc kubenswrapper[4894]: E1208 14:48:01.196842 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:01 crc kubenswrapper[4894]: E1208 14:48:01.196919 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:02 crc kubenswrapper[4894]: I1208 14:48:02.195923 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:02 crc kubenswrapper[4894]: I1208 14:48:02.195957 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:02 crc kubenswrapper[4894]: E1208 14:48:02.196374 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:02 crc kubenswrapper[4894]: E1208 14:48:02.196467 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:03 crc kubenswrapper[4894]: I1208 14:48:03.196862 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:03 crc kubenswrapper[4894]: E1208 14:48:03.197049 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:03 crc kubenswrapper[4894]: I1208 14:48:03.197138 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:03 crc kubenswrapper[4894]: E1208 14:48:03.197225 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:04 crc kubenswrapper[4894]: I1208 14:48:04.196261 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:04 crc kubenswrapper[4894]: E1208 14:48:04.196380 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:04 crc kubenswrapper[4894]: I1208 14:48:04.196596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:04 crc kubenswrapper[4894]: E1208 14:48:04.196775 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:05 crc kubenswrapper[4894]: I1208 14:48:05.196682 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:05 crc kubenswrapper[4894]: E1208 14:48:05.196800 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:05 crc kubenswrapper[4894]: I1208 14:48:05.197483 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:05 crc kubenswrapper[4894]: E1208 14:48:05.197730 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:06 crc kubenswrapper[4894]: I1208 14:48:06.196020 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:06 crc kubenswrapper[4894]: E1208 14:48:06.196160 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:06 crc kubenswrapper[4894]: I1208 14:48:06.196502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:06 crc kubenswrapper[4894]: E1208 14:48:06.196764 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:07 crc kubenswrapper[4894]: I1208 14:48:07.196317 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:07 crc kubenswrapper[4894]: I1208 14:48:07.196313 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:07 crc kubenswrapper[4894]: E1208 14:48:07.197432 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:07 crc kubenswrapper[4894]: E1208 14:48:07.197350 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:08 crc kubenswrapper[4894]: I1208 14:48:08.196142 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:08 crc kubenswrapper[4894]: I1208 14:48:08.196189 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:08 crc kubenswrapper[4894]: E1208 14:48:08.196266 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:08 crc kubenswrapper[4894]: E1208 14:48:08.196354 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:09 crc kubenswrapper[4894]: I1208 14:48:09.195846 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:09 crc kubenswrapper[4894]: I1208 14:48:09.195861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:09 crc kubenswrapper[4894]: E1208 14:48:09.198034 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:09 crc kubenswrapper[4894]: E1208 14:48:09.198183 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:09 crc kubenswrapper[4894]: I1208 14:48:09.199427 4894 scope.go:117] "RemoveContainer" containerID="d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189" Dec 08 14:48:09 crc kubenswrapper[4894]: E1208 14:48:09.199714 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c25bk_openshift-ovn-kubernetes(fb212a28-36c5-440e-8965-986352c5d3ea)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" Dec 08 14:48:10 crc kubenswrapper[4894]: I1208 14:48:10.196701 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:10 crc kubenswrapper[4894]: I1208 14:48:10.196752 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:10 crc kubenswrapper[4894]: E1208 14:48:10.196830 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:10 crc kubenswrapper[4894]: E1208 14:48:10.196894 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.196117 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.196197 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:11 crc kubenswrapper[4894]: E1208 14:48:11.196414 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:11 crc kubenswrapper[4894]: E1208 14:48:11.196546 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.763931 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/1.log" Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.764687 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/0.log" Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.764752 4894 generic.go:334] "Generic (PLEG): container finished" podID="be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235" containerID="8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937" exitCode=1 Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.764806 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerDied","Data":"8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937"} Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.764908 4894 scope.go:117] "RemoveContainer" containerID="6ae2aac8b6e28a0a1f2c239db2eaebae52109cc6f42792c08795d93a4ad6cf78" Dec 08 14:48:11 crc kubenswrapper[4894]: I1208 14:48:11.765494 4894 scope.go:117] "RemoveContainer" containerID="8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937" Dec 08 14:48:11 crc kubenswrapper[4894]: E1208 14:48:11.765741 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-jtx8z_openshift-multus(be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235)\"" pod="openshift-multus/multus-jtx8z" podUID="be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235" Dec 08 14:48:12 crc kubenswrapper[4894]: I1208 14:48:12.196118 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:12 crc kubenswrapper[4894]: E1208 14:48:12.196257 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:12 crc kubenswrapper[4894]: I1208 14:48:12.196435 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:12 crc kubenswrapper[4894]: E1208 14:48:12.196494 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:12 crc kubenswrapper[4894]: I1208 14:48:12.768670 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/1.log" Dec 08 14:48:13 crc kubenswrapper[4894]: I1208 14:48:13.196468 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:13 crc kubenswrapper[4894]: I1208 14:48:13.196593 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:13 crc kubenswrapper[4894]: E1208 14:48:13.196740 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:13 crc kubenswrapper[4894]: E1208 14:48:13.196941 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:14 crc kubenswrapper[4894]: I1208 14:48:14.196402 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:14 crc kubenswrapper[4894]: I1208 14:48:14.196454 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:14 crc kubenswrapper[4894]: E1208 14:48:14.196531 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:14 crc kubenswrapper[4894]: E1208 14:48:14.196598 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:15 crc kubenswrapper[4894]: I1208 14:48:15.196012 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:15 crc kubenswrapper[4894]: E1208 14:48:15.196188 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:15 crc kubenswrapper[4894]: I1208 14:48:15.196238 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:15 crc kubenswrapper[4894]: E1208 14:48:15.196455 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:16 crc kubenswrapper[4894]: I1208 14:48:16.196633 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:16 crc kubenswrapper[4894]: E1208 14:48:16.197034 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:16 crc kubenswrapper[4894]: I1208 14:48:16.196640 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:16 crc kubenswrapper[4894]: E1208 14:48:16.197225 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:17 crc kubenswrapper[4894]: I1208 14:48:17.196657 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:17 crc kubenswrapper[4894]: I1208 14:48:17.196704 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:17 crc kubenswrapper[4894]: E1208 14:48:17.196791 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:17 crc kubenswrapper[4894]: E1208 14:48:17.197014 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:18 crc kubenswrapper[4894]: I1208 14:48:18.196706 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:18 crc kubenswrapper[4894]: I1208 14:48:18.196740 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:18 crc kubenswrapper[4894]: E1208 14:48:18.196875 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:18 crc kubenswrapper[4894]: E1208 14:48:18.197220 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:19 crc kubenswrapper[4894]: E1208 14:48:19.154426 4894 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 08 14:48:19 crc kubenswrapper[4894]: I1208 14:48:19.196281 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:19 crc kubenswrapper[4894]: I1208 14:48:19.196317 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:19 crc kubenswrapper[4894]: E1208 14:48:19.198953 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:19 crc kubenswrapper[4894]: E1208 14:48:19.199148 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:19 crc kubenswrapper[4894]: E1208 14:48:19.300545 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 14:48:20 crc kubenswrapper[4894]: I1208 14:48:20.196563 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:20 crc kubenswrapper[4894]: I1208 14:48:20.196563 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:20 crc kubenswrapper[4894]: E1208 14:48:20.196752 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:20 crc kubenswrapper[4894]: E1208 14:48:20.196894 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:21 crc kubenswrapper[4894]: I1208 14:48:21.197118 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:21 crc kubenswrapper[4894]: E1208 14:48:21.197311 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:21 crc kubenswrapper[4894]: I1208 14:48:21.197390 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:21 crc kubenswrapper[4894]: E1208 14:48:21.197671 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:22 crc kubenswrapper[4894]: I1208 14:48:22.196153 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:22 crc kubenswrapper[4894]: I1208 14:48:22.196215 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:22 crc kubenswrapper[4894]: E1208 14:48:22.196257 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:22 crc kubenswrapper[4894]: E1208 14:48:22.196372 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.195861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:23 crc kubenswrapper[4894]: E1208 14:48:23.196309 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.196460 4894 scope.go:117] "RemoveContainer" containerID="d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189" Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.195904 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.196662 4894 scope.go:117] "RemoveContainer" containerID="8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937" Dec 08 14:48:23 crc kubenswrapper[4894]: E1208 14:48:23.199440 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.809282 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/1.log" Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.809680 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerStarted","Data":"49c32397a748d68b9ff9dc6f0fddb18ac746953d7dce6f7c93da63f6b3337ba7"} Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.811712 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/3.log" Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.814240 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerStarted","Data":"533baecbcd1f2894c4dd1141d33b4fb719b2d10aff57238aa5684d387568d9f6"} Dec 08 14:48:23 crc kubenswrapper[4894]: I1208 14:48:23.814661 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:48:24 crc kubenswrapper[4894]: I1208 14:48:24.122966 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podStartSLOduration=106.12294519 podStartE2EDuration="1m46.12294519s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:23.853465396 +0000 UTC m=+124.953471531" watchObservedRunningTime="2025-12-08 14:48:24.12294519 +0000 UTC m=+125.222951335" Dec 08 14:48:24 crc kubenswrapper[4894]: I1208 14:48:24.123935 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-z2zz6"] Dec 08 14:48:24 crc kubenswrapper[4894]: I1208 14:48:24.124039 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:24 crc kubenswrapper[4894]: E1208 14:48:24.124124 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:24 crc kubenswrapper[4894]: I1208 14:48:24.196462 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:24 crc kubenswrapper[4894]: E1208 14:48:24.196596 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:24 crc kubenswrapper[4894]: E1208 14:48:24.301554 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 08 14:48:25 crc kubenswrapper[4894]: I1208 14:48:25.196446 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:25 crc kubenswrapper[4894]: E1208 14:48:25.196845 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:25 crc kubenswrapper[4894]: I1208 14:48:25.196619 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:25 crc kubenswrapper[4894]: I1208 14:48:25.196543 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:25 crc kubenswrapper[4894]: E1208 14:48:25.196916 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:25 crc kubenswrapper[4894]: E1208 14:48:25.197055 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:26 crc kubenswrapper[4894]: I1208 14:48:26.196804 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:26 crc kubenswrapper[4894]: E1208 14:48:26.196948 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:27 crc kubenswrapper[4894]: I1208 14:48:27.196041 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:27 crc kubenswrapper[4894]: I1208 14:48:27.196052 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:27 crc kubenswrapper[4894]: E1208 14:48:27.196226 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:27 crc kubenswrapper[4894]: I1208 14:48:27.196260 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:27 crc kubenswrapper[4894]: E1208 14:48:27.196419 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:27 crc kubenswrapper[4894]: E1208 14:48:27.196498 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:28 crc kubenswrapper[4894]: I1208 14:48:28.196664 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:28 crc kubenswrapper[4894]: E1208 14:48:28.196785 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 08 14:48:29 crc kubenswrapper[4894]: I1208 14:48:29.196748 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:29 crc kubenswrapper[4894]: I1208 14:48:29.196747 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:29 crc kubenswrapper[4894]: I1208 14:48:29.196797 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:29 crc kubenswrapper[4894]: E1208 14:48:29.197659 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2zz6" podUID="49a05fc4-1361-4ebc-891a-e2b49df28ffa" Dec 08 14:48:29 crc kubenswrapper[4894]: E1208 14:48:29.197757 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 08 14:48:29 crc kubenswrapper[4894]: E1208 14:48:29.197903 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 08 14:48:30 crc kubenswrapper[4894]: I1208 14:48:30.196478 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:30 crc kubenswrapper[4894]: I1208 14:48:30.199108 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 08 14:48:30 crc kubenswrapper[4894]: I1208 14:48:30.200924 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 08 14:48:31 crc kubenswrapper[4894]: I1208 14:48:31.198193 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:31 crc kubenswrapper[4894]: I1208 14:48:31.198640 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:31 crc kubenswrapper[4894]: I1208 14:48:31.198905 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:48:31 crc kubenswrapper[4894]: I1208 14:48:31.200356 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 08 14:48:31 crc kubenswrapper[4894]: I1208 14:48:31.201146 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 08 14:48:31 crc kubenswrapper[4894]: I1208 14:48:31.202561 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 08 14:48:31 crc kubenswrapper[4894]: I1208 14:48:31.203780 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.928887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.963619 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-f9827"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.964138 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.964529 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.964983 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.965837 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.966302 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.967783 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.968240 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q6l7x"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.968795 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.969100 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-sq5fn"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.969202 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.970496 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.970977 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-955zh"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971029 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971082 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971105 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971203 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971251 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971263 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971035 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971322 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971279 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971688 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971729 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971831 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.971998 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.972142 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.973865 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7nbxt"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.974317 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-77css"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.974670 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.974685 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.974671 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.978160 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.978418 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.978511 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.978624 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.979196 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.980134 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.980184 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.980367 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.983864 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4scj2"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.984299 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.984383 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.991883 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.993277 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.998874 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9r5rc"] Dec 08 14:48:32 crc kubenswrapper[4894]: I1208 14:48:32.999548 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9r5rc" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.001664 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.002145 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.002460 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-99z5g"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.002774 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.013331 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.013692 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.013719 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.013830 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.013843 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.014251 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.014269 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.014453 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.014467 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.014668 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.016534 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.016806 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017270 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017287 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017409 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-auth-proxy-config\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017440 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh8rc\" (UniqueName: \"kubernetes.io/projected/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-kube-api-access-dh8rc\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017473 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-machine-approver-tls\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017511 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-config\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017601 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017666 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017701 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017787 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.017913 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.018016 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.018250 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.018374 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.018387 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.018484 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.018708 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.018945 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019069 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019224 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019324 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019627 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019657 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019693 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019334 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019763 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019835 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019915 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.019937 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.020002 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.020011 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.020025 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.020095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.020113 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.023194 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.023427 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.023643 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.023800 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.023835 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.023921 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.023952 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.024026 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.024221 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.024240 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.024234 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.024548 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.029006 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.029134 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.033373 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.034884 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.034952 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xlkww"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.035561 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.036091 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.036545 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.038901 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vnpl8"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.039320 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.039688 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.040035 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.040993 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.041236 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.041405 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.041947 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.041973 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042080 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042090 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042151 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042170 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042217 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042226 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042155 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042309 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042688 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.042856 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.043003 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.043038 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.043166 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.044913 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.056929 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.057370 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.058300 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.061447 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.061884 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.062862 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.063144 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.063337 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.065070 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.065586 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.065959 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.070575 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.082337 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.083257 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.084179 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.085654 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.086411 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xgp2c"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.087417 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.088128 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.093750 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-w5l67"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.099960 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t52r4"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.100111 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.100563 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.100839 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.100875 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.101158 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.101338 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-swwjr"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.101578 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.101702 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.101811 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.102138 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.107529 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.108217 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25w2m"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.108240 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.108901 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.111836 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.112596 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.112708 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.115903 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.117236 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120022 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-config\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120074 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-console-config\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120104 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/32954141-28e6-48a5-9c87-3d0923fa1afe-metrics-tls\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120125 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32954141-28e6-48a5-9c87-3d0923fa1afe-trusted-ca\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120175 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-audit\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120211 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-encryption-config\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120239 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-config\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120264 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-service-ca-bundle\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120288 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbdnc\" (UniqueName: \"kubernetes.io/projected/03c75f78-3a86-4b5e-8ac1-2a22f80da570-kube-api-access-bbdnc\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldw6c\" (UniqueName: \"kubernetes.io/projected/4e9041d3-396d-4331-8b21-82fbf8d7b118-kube-api-access-ldw6c\") pod \"downloads-7954f5f757-9r5rc\" (UID: \"4e9041d3-396d-4331-8b21-82fbf8d7b118\") " pod="openshift-console/downloads-7954f5f757-9r5rc" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120351 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de736b74-9e9d-4ede-acfc-28d64582b060-config\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120392 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/64ff7573-1aca-4e17-91f6-ec4f44c86505-metrics-tls\") pod \"dns-operator-744455d44c-99z5g\" (UID: \"64ff7573-1aca-4e17-91f6-ec4f44c86505\") " pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffc6f048-b577-48b6-829d-dee383f4aeb9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97gr2\" (UniqueName: \"kubernetes.io/projected/368e61a3-8284-46f3-8901-cdcad7a729cb-kube-api-access-97gr2\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120547 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-serving-cert\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120633 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgmnn\" (UniqueName: \"kubernetes.io/projected/de736b74-9e9d-4ede-acfc-28d64582b060-kube-api-access-fgmnn\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120705 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8vnm\" (UniqueName: \"kubernetes.io/projected/22a82f00-b781-4068-b398-0b4c738db413-kube-api-access-f8vnm\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120744 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-config\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120863 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120888 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbb797fc-bf27-449b-b0c2-d72d29fc7845-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120916 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-auth-proxy-config\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120940 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de736b74-9e9d-4ede-acfc-28d64582b060-trusted-ca\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120961 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-etcd-client\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.120983 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-image-import-ca\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121007 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ckb4\" (UniqueName: \"kubernetes.io/projected/4231690c-869c-4813-abc1-4aca616a07c6-kube-api-access-4ckb4\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121032 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb797fc-bf27-449b-b0c2-d72d29fc7845-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121065 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-serving-cert\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121086 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-ca\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121115 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-oauth-config\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121134 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvtkb\" (UniqueName: \"kubernetes.io/projected/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-kube-api-access-gvtkb\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121166 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4231690c-869c-4813-abc1-4aca616a07c6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121207 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-trusted-ca-bundle\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121239 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-serving-cert\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de736b74-9e9d-4ede-acfc-28d64582b060-serving-cert\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121299 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-serving-cert\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121321 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvpgd\" (UniqueName: \"kubernetes.io/projected/64ff7573-1aca-4e17-91f6-ec4f44c86505-kube-api-access-dvpgd\") pod \"dns-operator-744455d44c-99z5g\" (UID: \"64ff7573-1aca-4e17-91f6-ec4f44c86505\") " pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121342 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffc6f048-b577-48b6-829d-dee383f4aeb9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121363 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnghf\" (UniqueName: \"kubernetes.io/projected/ffc6f048-b577-48b6-829d-dee383f4aeb9-kube-api-access-qnghf\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8j9k\" (UniqueName: \"kubernetes.io/projected/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-kube-api-access-t8j9k\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4231690c-869c-4813-abc1-4aca616a07c6-serving-cert\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121427 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121445 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-auth-proxy-config\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121449 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-oauth-serving-cert\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121513 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-config\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121534 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22a82f00-b781-4068-b398-0b4c738db413-config\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121568 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-etcd-serving-ca\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121592 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-client\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121613 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22a82f00-b781-4068-b398-0b4c738db413-images\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121701 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a82f00-b781-4068-b398-0b4c738db413-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121726 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03c75f78-3a86-4b5e-8ac1-2a22f80da570-audit-dir\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121756 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/32954141-28e6-48a5-9c87-3d0923fa1afe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121790 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03c75f78-3a86-4b5e-8ac1-2a22f80da570-node-pullsecrets\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.121941 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh8rc\" (UniqueName: \"kubernetes.io/projected/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-kube-api-access-dh8rc\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.122004 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-machine-approver-tls\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.122054 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb797fc-bf27-449b-b0c2-d72d29fc7845-config\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.122155 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-config\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.122305 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-service-ca\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.122340 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwzd8\" (UniqueName: \"kubernetes.io/projected/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-kube-api-access-jwzd8\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.122365 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-service-ca\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.122620 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t96g4\" (UniqueName: \"kubernetes.io/projected/32954141-28e6-48a5-9c87-3d0923fa1afe-kube-api-access-t96g4\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.125296 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.125378 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h6mkt"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.126037 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.128945 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q6l7x"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.129637 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.130794 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.130949 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.138697 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-machine-approver-tls\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.141407 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.141781 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-sq5fn"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.142941 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.143630 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-99z5g"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.144944 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-77css"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.147093 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.147127 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-955zh"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.147615 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.148661 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7nbxt"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.150732 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.166502 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.169081 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.174914 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.175976 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.177369 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.177527 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.178500 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.179854 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t52r4"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.180931 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rbk4g"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.182562 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.182662 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.183722 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.184890 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xgp2c"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.186021 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.186262 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.187288 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4scj2"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.188354 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xlkww"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.189427 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.190558 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9r5rc"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.191832 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.192968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.194097 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-bmrff"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.194710 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.195371 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vnpl8"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.200239 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-tmk6k"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.200881 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.200901 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rbk4g"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.200915 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-swwjr"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.200999 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.201285 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25w2m"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.202236 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.204394 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h6mkt"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.205664 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.205791 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.206799 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bmrff"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.208157 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tmk6k"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.209486 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.211034 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-297hf"] Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.212001 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223143 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbb797fc-bf27-449b-b0c2-d72d29fc7845-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223172 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de736b74-9e9d-4ede-acfc-28d64582b060-trusted-ca\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-etcd-client\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223214 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snqpl\" (UniqueName: \"kubernetes.io/projected/0872ad57-8232-471e-a68b-d7550c070794-kube-api-access-snqpl\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223231 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-client-ca\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223246 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-image-import-ca\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223261 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ckb4\" (UniqueName: \"kubernetes.io/projected/4231690c-869c-4813-abc1-4aca616a07c6-kube-api-access-4ckb4\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223279 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-etcd-client\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223296 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb797fc-bf27-449b-b0c2-d72d29fc7845-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223312 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59wbv\" (UniqueName: \"kubernetes.io/projected/8a9359ef-713b-449b-a468-78b54ccf6c64-kube-api-access-59wbv\") pod \"cluster-samples-operator-665b6dd947-9z9w9\" (UID: \"8a9359ef-713b-449b-a468-78b54ccf6c64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223339 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-serving-cert\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223354 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-ca\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223369 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/585fadc5-eeab-4236-b79e-9873939be9e0-tmpfs\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223392 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4kbv\" (UniqueName: \"kubernetes.io/projected/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-kube-api-access-q4kbv\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-oauth-config\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223617 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvtkb\" (UniqueName: \"kubernetes.io/projected/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-kube-api-access-gvtkb\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223633 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4231690c-869c-4813-abc1-4aca616a07c6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223651 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a9359ef-713b-449b-a468-78b54ccf6c64-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9z9w9\" (UID: \"8a9359ef-713b-449b-a468-78b54ccf6c64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223667 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-trusted-ca-bundle\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223687 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-serving-cert\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223702 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de736b74-9e9d-4ede-acfc-28d64582b060-serving-cert\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223719 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76jc2\" (UniqueName: \"kubernetes.io/projected/585fadc5-eeab-4236-b79e-9873939be9e0-kube-api-access-76jc2\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223738 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0872ad57-8232-471e-a68b-d7550c070794-audit-dir\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223753 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-serving-cert\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223768 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvpgd\" (UniqueName: \"kubernetes.io/projected/64ff7573-1aca-4e17-91f6-ec4f44c86505-kube-api-access-dvpgd\") pod \"dns-operator-744455d44c-99z5g\" (UID: \"64ff7573-1aca-4e17-91f6-ec4f44c86505\") " pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffc6f048-b577-48b6-829d-dee383f4aeb9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnghf\" (UniqueName: \"kubernetes.io/projected/ffc6f048-b577-48b6-829d-dee383f4aeb9-kube-api-access-qnghf\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8j9k\" (UniqueName: \"kubernetes.io/projected/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-kube-api-access-t8j9k\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223849 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4231690c-869c-4813-abc1-4aca616a07c6-serving-cert\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223865 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-serving-cert\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223881 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-proxy-tls\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223896 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68be7440-39c0-4161-b248-7bb4f78fa41b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223934 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223950 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/585fadc5-eeab-4236-b79e-9873939be9e0-webhook-cert\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223965 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-oauth-serving-cert\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223979 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-config\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.223996 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22a82f00-b781-4068-b398-0b4c738db413-config\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224013 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvn5v\" (UniqueName: \"kubernetes.io/projected/552313d0-57ae-4e50-ae23-d7a568e1afcd-kube-api-access-dvn5v\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-etcd-serving-ca\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224041 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-client\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224056 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22a82f00-b781-4068-b398-0b4c738db413-images\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224072 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224087 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224107 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxn96\" (UniqueName: \"kubernetes.io/projected/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-kube-api-access-pxn96\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a82f00-b781-4068-b398-0b4c738db413-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03c75f78-3a86-4b5e-8ac1-2a22f80da570-audit-dir\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224157 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/32954141-28e6-48a5-9c87-3d0923fa1afe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224172 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03c75f78-3a86-4b5e-8ac1-2a22f80da570-node-pullsecrets\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224186 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224208 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-encryption-config\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224223 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb797fc-bf27-449b-b0c2-d72d29fc7845-config\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224239 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-config\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6wz\" (UniqueName: \"kubernetes.io/projected/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-kube-api-access-5n6wz\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224286 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-service-ca\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224301 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwzd8\" (UniqueName: \"kubernetes.io/projected/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-kube-api-access-jwzd8\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224316 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-service-ca\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224333 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qr7n\" (UniqueName: \"kubernetes.io/projected/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-kube-api-access-6qr7n\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-default-certificate\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224366 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t96g4\" (UniqueName: \"kubernetes.io/projected/32954141-28e6-48a5-9c87-3d0923fa1afe-kube-api-access-t96g4\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224381 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-audit-policies\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224395 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-srv-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-console-config\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224424 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/32954141-28e6-48a5-9c87-3d0923fa1afe-metrics-tls\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224438 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32954141-28e6-48a5-9c87-3d0923fa1afe-trusted-ca\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224455 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-service-ca-bundle\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-audit\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224485 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-encryption-config\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224480 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-image-import-ca\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224502 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-config\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224539 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-service-ca-bundle\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224564 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224587 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbdnc\" (UniqueName: \"kubernetes.io/projected/03c75f78-3a86-4b5e-8ac1-2a22f80da570-kube-api-access-bbdnc\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224622 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-stats-auth\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224643 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldw6c\" (UniqueName: \"kubernetes.io/projected/4e9041d3-396d-4331-8b21-82fbf8d7b118-kube-api-access-ldw6c\") pod \"downloads-7954f5f757-9r5rc\" (UID: \"4e9041d3-396d-4331-8b21-82fbf8d7b118\") " pod="openshift-console/downloads-7954f5f757-9r5rc" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224662 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de736b74-9e9d-4ede-acfc-28d64582b060-config\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224679 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/64ff7573-1aca-4e17-91f6-ec4f44c86505-metrics-tls\") pod \"dns-operator-744455d44c-99z5g\" (UID: \"64ff7573-1aca-4e17-91f6-ec4f44c86505\") " pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224695 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffc6f048-b577-48b6-829d-dee383f4aeb9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224717 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224737 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552313d0-57ae-4e50-ae23-d7a568e1afcd-serving-cert\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224758 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97gr2\" (UniqueName: \"kubernetes.io/projected/368e61a3-8284-46f3-8901-cdcad7a729cb-kube-api-access-97gr2\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224774 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-serving-cert\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224809 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgmnn\" (UniqueName: \"kubernetes.io/projected/de736b74-9e9d-4ede-acfc-28d64582b060-kube-api-access-fgmnn\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224858 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8vnm\" (UniqueName: \"kubernetes.io/projected/22a82f00-b781-4068-b398-0b4c738db413-kube-api-access-f8vnm\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224882 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224897 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-config\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224913 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-metrics-certs\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-config\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224928 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68be7440-39c0-4161-b248-7bb4f78fa41b-config\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68be7440-39c0-4161-b248-7bb4f78fa41b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224972 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224988 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-images\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.225002 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/585fadc5-eeab-4236-b79e-9873939be9e0-apiservice-cert\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.225466 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-ca\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.225715 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-service-ca-bundle\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.224539 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de736b74-9e9d-4ede-acfc-28d64582b060-trusted-ca\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.226524 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-etcd-client\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.226590 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de736b74-9e9d-4ede-acfc-28d64582b060-config\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.226792 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-serving-cert\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.227028 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbb797fc-bf27-449b-b0c2-d72d29fc7845-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.227171 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03c75f78-3a86-4b5e-8ac1-2a22f80da570-audit-dir\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.227389 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03c75f78-3a86-4b5e-8ac1-2a22f80da570-node-pullsecrets\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.227715 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-service-ca\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.227881 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4231690c-869c-4813-abc1-4aca616a07c6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.228339 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb797fc-bf27-449b-b0c2-d72d29fc7845-config\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.228662 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.228927 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-config\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.229007 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-trusted-ca-bundle\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.229081 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/64ff7573-1aca-4e17-91f6-ec4f44c86505-metrics-tls\") pod \"dns-operator-744455d44c-99z5g\" (UID: \"64ff7573-1aca-4e17-91f6-ec4f44c86505\") " pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.229193 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.229677 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-service-ca\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.229713 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-serving-cert\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.229787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.230386 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-console-config\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.230440 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/22a82f00-b781-4068-b398-0b4c738db413-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.230892 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32954141-28e6-48a5-9c87-3d0923fa1afe-trusted-ca\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.231073 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-serving-cert\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.231136 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-oauth-serving-cert\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.231810 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-audit\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.232015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22a82f00-b781-4068-b398-0b4c738db413-config\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.232207 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/22a82f00-b781-4068-b398-0b4c738db413-images\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.232601 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ffc6f048-b577-48b6-829d-dee383f4aeb9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.232728 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-trusted-ca-bundle\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.232936 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-config\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.232974 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03c75f78-3a86-4b5e-8ac1-2a22f80da570-etcd-serving-ca\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.233044 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4231690c-869c-4813-abc1-4aca616a07c6-serving-cert\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.233408 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/32954141-28e6-48a5-9c87-3d0923fa1afe-metrics-tls\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.233516 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-etcd-client\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.233679 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-serving-cert\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.234568 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.235897 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de736b74-9e9d-4ede-acfc-28d64582b060-serving-cert\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.236413 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03c75f78-3a86-4b5e-8ac1-2a22f80da570-encryption-config\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.241467 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-oauth-config\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.250312 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.251073 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffc6f048-b577-48b6-829d-dee383f4aeb9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.269253 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.286430 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.307581 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325550 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552313d0-57ae-4e50-ae23-d7a568e1afcd-serving-cert\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325586 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325611 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325638 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325653 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-config\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325682 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68be7440-39c0-4161-b248-7bb4f78fa41b-config\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325700 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68be7440-39c0-4161-b248-7bb4f78fa41b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325718 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-metrics-certs\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325734 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-images\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325749 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/585fadc5-eeab-4236-b79e-9873939be9e0-apiservice-cert\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325767 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snqpl\" (UniqueName: \"kubernetes.io/projected/0872ad57-8232-471e-a68b-d7550c070794-kube-api-access-snqpl\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-client-ca\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325799 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-etcd-client\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325859 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59wbv\" (UniqueName: \"kubernetes.io/projected/8a9359ef-713b-449b-a468-78b54ccf6c64-kube-api-access-59wbv\") pod \"cluster-samples-operator-665b6dd947-9z9w9\" (UID: \"8a9359ef-713b-449b-a468-78b54ccf6c64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325883 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/585fadc5-eeab-4236-b79e-9873939be9e0-tmpfs\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325908 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4kbv\" (UniqueName: \"kubernetes.io/projected/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-kube-api-access-q4kbv\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325924 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a9359ef-713b-449b-a468-78b54ccf6c64-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9z9w9\" (UID: \"8a9359ef-713b-449b-a468-78b54ccf6c64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325945 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0872ad57-8232-471e-a68b-d7550c070794-audit-dir\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325960 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76jc2\" (UniqueName: \"kubernetes.io/projected/585fadc5-eeab-4236-b79e-9873939be9e0-kube-api-access-76jc2\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.325992 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-proxy-tls\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68be7440-39c0-4161-b248-7bb4f78fa41b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326026 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-serving-cert\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326041 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/585fadc5-eeab-4236-b79e-9873939be9e0-webhook-cert\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326063 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvn5v\" (UniqueName: \"kubernetes.io/projected/552313d0-57ae-4e50-ae23-d7a568e1afcd-kube-api-access-dvn5v\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326097 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxn96\" (UniqueName: \"kubernetes.io/projected/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-kube-api-access-pxn96\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326148 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-encryption-config\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326169 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n6wz\" (UniqueName: \"kubernetes.io/projected/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-kube-api-access-5n6wz\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326233 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-audit-policies\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326252 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qr7n\" (UniqueName: \"kubernetes.io/projected/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-kube-api-access-6qr7n\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326255 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0872ad57-8232-471e-a68b-d7550c070794-audit-dir\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326268 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-default-certificate\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326362 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-srv-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326429 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-service-ca-bundle\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326518 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326555 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326576 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-stats-auth\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326652 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/585fadc5-eeab-4236-b79e-9873939be9e0-tmpfs\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.326691 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.327191 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0872ad57-8232-471e-a68b-d7550c070794-audit-policies\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.327392 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.329441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-serving-cert\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.329674 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-encryption-config\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.329860 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8a9359ef-713b-449b-a468-78b54ccf6c64-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9z9w9\" (UID: \"8a9359ef-713b-449b-a468-78b54ccf6c64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.332235 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.332393 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552313d0-57ae-4e50-ae23-d7a568e1afcd-serving-cert\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.332746 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0872ad57-8232-471e-a68b-d7550c070794-etcd-client\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.346253 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.366518 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.386044 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.406308 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.407603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-config\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.425997 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.436852 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-client-ca\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.446037 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.466336 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.490727 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.505996 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.526131 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.546594 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.567173 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.586564 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.606810 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.626981 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.645791 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.666212 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.686443 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.706153 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.726098 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.737369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-images\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.746508 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.766189 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.769416 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68be7440-39c0-4161-b248-7bb4f78fa41b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.786348 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.796651 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68be7440-39c0-4161-b248-7bb4f78fa41b-config\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.806741 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.827692 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.842127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-proxy-tls\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.867145 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.892452 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.907467 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.927651 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.947267 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.967368 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 08 14:48:33 crc kubenswrapper[4894]: I1208 14:48:33.987050 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.007268 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.026742 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.046729 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.060938 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-default-certificate\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.066892 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.085194 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-stats-auth\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.087074 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.099411 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-metrics-certs\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.105022 4894 request.go:700] Waited for 1.004460245s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.107701 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.128214 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.138753 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-service-ca-bundle\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.147665 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.166705 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.187203 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.206412 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.226990 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.240445 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/585fadc5-eeab-4236-b79e-9873939be9e0-webhook-cert\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.241283 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/585fadc5-eeab-4236-b79e-9873939be9e0-apiservice-cert\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.247717 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.266659 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.286107 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.306180 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 08 14:48:34 crc kubenswrapper[4894]: E1208 14:48:34.326672 4894 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Dec 08 14:48:34 crc kubenswrapper[4894]: E1208 14:48:34.326757 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-profile-collector-cert podName:e6badbf9-7e6b-4c55-a56a-a7153faae4e1 nodeName:}" failed. No retries permitted until 2025-12-08 14:48:34.826735677 +0000 UTC m=+135.926741792 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-profile-collector-cert") pod "olm-operator-6b444d44fb-hnsbl" (UID: "e6badbf9-7e6b-4c55-a56a-a7153faae4e1") : failed to sync secret cache: timed out waiting for the condition Dec 08 14:48:34 crc kubenswrapper[4894]: E1208 14:48:34.326793 4894 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 08 14:48:34 crc kubenswrapper[4894]: E1208 14:48:34.327013 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-srv-cert podName:e6badbf9-7e6b-4c55-a56a-a7153faae4e1 nodeName:}" failed. No retries permitted until 2025-12-08 14:48:34.826926944 +0000 UTC m=+135.926933069 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-srv-cert") pod "olm-operator-6b444d44fb-hnsbl" (UID: "e6badbf9-7e6b-4c55-a56a-a7153faae4e1") : failed to sync secret cache: timed out waiting for the condition Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.327325 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.346057 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.366414 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.386407 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.406141 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.427550 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.447104 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.466485 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.486830 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.506764 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.534613 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.546958 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.566695 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.622275 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh8rc\" (UniqueName: \"kubernetes.io/projected/f23bacbf-0938-4ecc-bbce-b816bad0a9e0-kube-api-access-dh8rc\") pod \"machine-approver-56656f9798-f9827\" (UID: \"f23bacbf-0938-4ecc-bbce-b816bad0a9e0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.626853 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.646247 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.666537 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.686974 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.706606 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.726743 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.747150 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.766753 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.787379 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.794000 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.811060 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.830270 4894 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.847033 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.847182 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.847289 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-srv-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.850594 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-srv-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.850766 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.853130 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" event={"ID":"f23bacbf-0938-4ecc-bbce-b816bad0a9e0","Type":"ContainerStarted","Data":"01e3002f7b55df416171333cc894f6a8f16c753f95e77075e4b2a2c656b55148"} Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.867275 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.887238 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.906272 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.925770 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.948210 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.966608 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 08 14:48:34 crc kubenswrapper[4894]: I1208 14:48:34.986445 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.006764 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.027742 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.046023 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.080357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ckb4\" (UniqueName: \"kubernetes.io/projected/4231690c-869c-4813-abc1-4aca616a07c6-kube-api-access-4ckb4\") pod \"openshift-config-operator-7777fb866f-9bx4w\" (UID: \"4231690c-869c-4813-abc1-4aca616a07c6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.099188 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvtkb\" (UniqueName: \"kubernetes.io/projected/f2bb84da-9788-42b9-8c65-900ce3b2b7aa-kube-api-access-gvtkb\") pod \"openshift-controller-manager-operator-756b6f6bc6-vx8xs\" (UID: \"f2bb84da-9788-42b9-8c65-900ce3b2b7aa\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.105094 4894 request.go:700] Waited for 1.879941417s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/serviceaccounts/kube-apiserver-operator/token Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.120630 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbb797fc-bf27-449b-b0c2-d72d29fc7845-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kwjpt\" (UID: \"cbb797fc-bf27-449b-b0c2-d72d29fc7845\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.140250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbdnc\" (UniqueName: \"kubernetes.io/projected/03c75f78-3a86-4b5e-8ac1-2a22f80da570-kube-api-access-bbdnc\") pod \"apiserver-76f77b778f-q6l7x\" (UID: \"03c75f78-3a86-4b5e-8ac1-2a22f80da570\") " pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.160941 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldw6c\" (UniqueName: \"kubernetes.io/projected/4e9041d3-396d-4331-8b21-82fbf8d7b118-kube-api-access-ldw6c\") pod \"downloads-7954f5f757-9r5rc\" (UID: \"4e9041d3-396d-4331-8b21-82fbf8d7b118\") " pod="openshift-console/downloads-7954f5f757-9r5rc" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.188218 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.199713 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.202749 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97gr2\" (UniqueName: \"kubernetes.io/projected/368e61a3-8284-46f3-8901-cdcad7a729cb-kube-api-access-97gr2\") pod \"console-f9d7485db-77css\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.219266 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvpgd\" (UniqueName: \"kubernetes.io/projected/64ff7573-1aca-4e17-91f6-ec4f44c86505-kube-api-access-dvpgd\") pod \"dns-operator-744455d44c-99z5g\" (UID: \"64ff7573-1aca-4e17-91f6-ec4f44c86505\") " pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.239526 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/32954141-28e6-48a5-9c87-3d0923fa1afe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.259683 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwzd8\" (UniqueName: \"kubernetes.io/projected/2f0d52f8-ca4e-4062-9d57-ff54c976d49e-kube-api-access-jwzd8\") pod \"authentication-operator-69f744f599-sq5fn\" (UID: \"2f0d52f8-ca4e-4062-9d57-ff54c976d49e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.263568 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.280239 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgmnn\" (UniqueName: \"kubernetes.io/projected/de736b74-9e9d-4ede-acfc-28d64582b060-kube-api-access-fgmnn\") pod \"console-operator-58897d9998-955zh\" (UID: \"de736b74-9e9d-4ede-acfc-28d64582b060\") " pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.286153 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.313897 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.321180 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8j9k\" (UniqueName: \"kubernetes.io/projected/65ac3b9e-19fd-4ae8-8c05-bb566ca43be4-kube-api-access-t8j9k\") pod \"etcd-operator-b45778765-4scj2\" (UID: \"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.321394 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9r5rc" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.340067 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t96g4\" (UniqueName: \"kubernetes.io/projected/32954141-28e6-48a5-9c87-3d0923fa1afe-kube-api-access-t96g4\") pod \"ingress-operator-5b745b69d9-tlvsl\" (UID: \"32954141-28e6-48a5-9c87-3d0923fa1afe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.360015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68be7440-39c0-4161-b248-7bb4f78fa41b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-75fl8\" (UID: \"68be7440-39c0-4161-b248-7bb4f78fa41b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.368623 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.380475 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59wbv\" (UniqueName: \"kubernetes.io/projected/8a9359ef-713b-449b-a468-78b54ccf6c64-kube-api-access-59wbv\") pod \"cluster-samples-operator-665b6dd947-9z9w9\" (UID: \"8a9359ef-713b-449b-a468-78b54ccf6c64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.399725 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snqpl\" (UniqueName: \"kubernetes.io/projected/0872ad57-8232-471e-a68b-d7550c070794-kube-api-access-snqpl\") pod \"apiserver-7bbb656c7d-l2fzd\" (UID: \"0872ad57-8232-471e-a68b-d7550c070794\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.418739 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvn5v\" (UniqueName: \"kubernetes.io/projected/552313d0-57ae-4e50-ae23-d7a568e1afcd-kube-api-access-dvn5v\") pod \"route-controller-manager-6576b87f9c-txhpt\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.438664 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4kbv\" (UniqueName: \"kubernetes.io/projected/e00f1298-2c1d-4cdd-825f-15f8b9809e0d-kube-api-access-q4kbv\") pod \"openshift-apiserver-operator-796bbdcf4f-n2zms\" (UID: \"e00f1298-2c1d-4cdd-825f-15f8b9809e0d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.459097 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76jc2\" (UniqueName: \"kubernetes.io/projected/585fadc5-eeab-4236-b79e-9873939be9e0-kube-api-access-76jc2\") pod \"packageserver-d55dfcdfc-td4j4\" (UID: \"585fadc5-eeab-4236-b79e-9873939be9e0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.459391 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.480623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxn96\" (UniqueName: \"kubernetes.io/projected/e6badbf9-7e6b-4c55-a56a-a7153faae4e1-kube-api-access-pxn96\") pod \"olm-operator-6b444d44fb-hnsbl\" (UID: \"e6badbf9-7e6b-4c55-a56a-a7153faae4e1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.494653 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.512964 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.517224 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.520357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qr7n\" (UniqueName: \"kubernetes.io/projected/4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e-kube-api-access-6qr7n\") pod \"machine-config-operator-74547568cd-mvkpm\" (UID: \"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:35 crc kubenswrapper[4894]: I1208 14:48:35.529939 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.151337 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnghf\" (UniqueName: \"kubernetes.io/projected/ffc6f048-b577-48b6-829d-dee383f4aeb9-kube-api-access-qnghf\") pod \"kube-storage-version-migrator-operator-b67b599dd-p76j7\" (UID: \"ffc6f048-b577-48b6-829d-dee383f4aeb9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.151689 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.151857 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.151994 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.152006 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.152111 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.152192 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.152306 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.152295 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153088 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153151 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a9ef065-72ff-4aa4-a729-c4e9884728e6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153174 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-certificates\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153192 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-trusted-ca\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153214 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a9ef065-72ff-4aa4-a729-c4e9884728e6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153230 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk8qv\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-kube-api-access-sk8qv\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153252 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvsv2\" (UniqueName: \"kubernetes.io/projected/8d36c61e-d514-4b44-a246-dd24d56c47dd-kube-api-access-lvsv2\") pod \"migrator-59844c95c7-h66ct\" (UID: \"8d36c61e-d514-4b44-a246-dd24d56c47dd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153330 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-bound-sa-token\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153370 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-tls\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.153910 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8vnm\" (UniqueName: \"kubernetes.io/projected/22a82f00-b781-4068-b398-0b4c738db413-kube-api-access-f8vnm\") pod \"machine-api-operator-5694c8668f-7nbxt\" (UID: \"22a82f00-b781-4068-b398-0b4c738db413\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.154246 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:36.654230599 +0000 UTC m=+137.754236774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.160525 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n6wz\" (UniqueName: \"kubernetes.io/projected/37ec9532-4218-4e18-bdb2-d9e8781f7cb7-kube-api-access-5n6wz\") pod \"router-default-5444994796-w5l67\" (UID: \"37ec9532-4218-4e18-bdb2-d9e8781f7cb7\") " pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.198142 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.254694 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.255230 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:36.755204456 +0000 UTC m=+137.855210571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.256472 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-trusted-ca\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.257365 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.257458 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1262f8ec-a2fc-4b25-b15e-447c35472b6d-profile-collector-cert\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.257543 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.258346 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5a424e2c-962d-444a-a5a3-57ef0e6133e9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t52r4\" (UID: \"5a424e2c-962d-444a-a5a3-57ef0e6133e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.258485 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2kkb\" (UniqueName: \"kubernetes.io/projected/0dff8795-f189-46d1-8d8b-aae4dad68c63-kube-api-access-m2kkb\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.258547 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.258635 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e899bcbb-0ffd-4989-bba4-47b61fe07832-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-njrj9\" (UID: \"e899bcbb-0ffd-4989-bba4-47b61fe07832\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.258955 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-trusted-ca\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.259018 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-bound-sa-token\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.259377 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.259579 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/774d8fe8-cc03-414c-a953-2275e5fd5123-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.259784 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/774d8fe8-cc03-414c-a953-2275e5fd5123-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.260724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dff8795-f189-46d1-8d8b-aae4dad68c63-serving-cert\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261378 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-serving-cert\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261456 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-client-ca\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261584 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261628 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a9ef065-72ff-4aa4-a729-c4e9884728e6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261722 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261772 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhhkx\" (UniqueName: \"kubernetes.io/projected/e899bcbb-0ffd-4989-bba4-47b61fe07832-kube-api-access-nhhkx\") pod \"control-plane-machine-set-operator-78cbb6b69f-njrj9\" (UID: \"e899bcbb-0ffd-4989-bba4-47b61fe07832\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261793 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261809 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn745\" (UniqueName: \"kubernetes.io/projected/1262f8ec-a2fc-4b25-b15e-447c35472b6d-kube-api-access-hn745\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.261849 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262300 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhzp\" (UniqueName: \"kubernetes.io/projected/774d8fe8-cc03-414c-a953-2275e5fd5123-kube-api-access-dkhzp\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-config\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a9ef065-72ff-4aa4-a729-c4e9884728e6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262404 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk8qv\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-kube-api-access-sk8qv\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262426 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262448 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xhcj\" (UniqueName: \"kubernetes.io/projected/d2838428-72ae-469d-884d-6660de5ce2f1-kube-api-access-2xhcj\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262495 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8cdd7414-21ba-4649-9514-66f2e3688022-audit-dir\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262519 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvsv2\" (UniqueName: \"kubernetes.io/projected/8d36c61e-d514-4b44-a246-dd24d56c47dd-kube-api-access-lvsv2\") pod \"migrator-59844c95c7-h66ct\" (UID: \"8d36c61e-d514-4b44-a246-dd24d56c47dd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262613 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262649 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfwdc\" (UniqueName: \"kubernetes.io/projected/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-kube-api-access-zfwdc\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.262255 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a9ef065-72ff-4aa4-a729-c4e9884728e6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264134 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f749021-69b6-4051-bc44-49d2334f087b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q4dmf\" (UID: \"1f749021-69b6-4051-bc44-49d2334f087b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264163 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264186 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264206 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/774d8fe8-cc03-414c-a953-2275e5fd5123-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264239 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7pcz\" (UniqueName: \"kubernetes.io/projected/5a424e2c-962d-444a-a5a3-57ef0e6133e9-kube-api-access-m7pcz\") pod \"multus-admission-controller-857f4d67dd-t52r4\" (UID: \"5a424e2c-962d-444a-a5a3-57ef0e6133e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264261 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-tls\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264329 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264352 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-config\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h28nv\" (UniqueName: \"kubernetes.io/projected/8cdd7414-21ba-4649-9514-66f2e3688022-kube-api-access-h28nv\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264414 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264517 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264546 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfx96\" (UniqueName: \"kubernetes.io/projected/1f749021-69b6-4051-bc44-49d2334f087b-kube-api-access-xfx96\") pod \"package-server-manager-789f6589d5-q4dmf\" (UID: \"1f749021-69b6-4051-bc44-49d2334f087b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264622 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1262f8ec-a2fc-4b25-b15e-447c35472b6d-srv-cert\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264644 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-certificates\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264711 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-audit-policies\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.264750 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.267167 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:36.76714984 +0000 UTC m=+137.867155955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.268576 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-certificates\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.270492 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-tls\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.275397 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a9ef065-72ff-4aa4-a729-c4e9884728e6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.302351 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-bound-sa-token\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.333936 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvsv2\" (UniqueName: \"kubernetes.io/projected/8d36c61e-d514-4b44-a246-dd24d56c47dd-kube-api-access-lvsv2\") pod \"migrator-59844c95c7-h66ct\" (UID: \"8d36c61e-d514-4b44-a246-dd24d56c47dd\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.359654 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk8qv\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-kube-api-access-sk8qv\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365271 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.365394 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:36.865370564 +0000 UTC m=+137.965376679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365477 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365510 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfx96\" (UniqueName: \"kubernetes.io/projected/1f749021-69b6-4051-bc44-49d2334f087b-kube-api-access-xfx96\") pod \"package-server-manager-789f6589d5-q4dmf\" (UID: \"1f749021-69b6-4051-bc44-49d2334f087b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365533 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1262f8ec-a2fc-4b25-b15e-447c35472b6d-srv-cert\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-audit-policies\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365630 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365658 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pw4c\" (UniqueName: \"kubernetes.io/projected/b6dbcd2c-e683-4318-860b-447c2df78fdc-kube-api-access-8pw4c\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365694 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1262f8ec-a2fc-4b25-b15e-447c35472b6d-profile-collector-cert\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365719 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365743 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4dgs\" (UniqueName: \"kubernetes.io/projected/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-kube-api-access-d4dgs\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnw7n\" (UniqueName: \"kubernetes.io/projected/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-kube-api-access-nnw7n\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.365824 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:36.865804718 +0000 UTC m=+137.965810913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365854 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb6t5\" (UniqueName: \"kubernetes.io/projected/34716cea-437d-4d43-86e6-0bdd54bb7ce8-kube-api-access-rb6t5\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365917 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-registration-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.365942 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkts2\" (UniqueName: \"kubernetes.io/projected/cc98c09a-446c-45ab-9f91-2f558c153059-kube-api-access-rkts2\") pod \"ingress-canary-bmrff\" (UID: \"cc98c09a-446c-45ab-9f91-2f558c153059\") " pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366000 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366022 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5a424e2c-962d-444a-a5a3-57ef0e6133e9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t52r4\" (UID: \"5a424e2c-962d-444a-a5a3-57ef0e6133e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366038 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2kkb\" (UniqueName: \"kubernetes.io/projected/0dff8795-f189-46d1-8d8b-aae4dad68c63-kube-api-access-m2kkb\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366062 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc98c09a-446c-45ab-9f91-2f558c153059-cert\") pod \"ingress-canary-bmrff\" (UID: \"cc98c09a-446c-45ab-9f91-2f558c153059\") " pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e899bcbb-0ffd-4989-bba4-47b61fe07832-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-njrj9\" (UID: \"e899bcbb-0ffd-4989-bba4-47b61fe07832\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/34716cea-437d-4d43-86e6-0bdd54bb7ce8-signing-cabundle\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366118 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/34716cea-437d-4d43-86e6-0bdd54bb7ce8-signing-key\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366144 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-node-bootstrap-token\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366162 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-mountpoint-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366182 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/774d8fe8-cc03-414c-a953-2275e5fd5123-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366218 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6dbcd2c-e683-4318-860b-447c2df78fdc-secret-volume\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366235 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/774d8fe8-cc03-414c-a953-2275e5fd5123-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366254 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-metrics-tls\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366293 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dff8795-f189-46d1-8d8b-aae4dad68c63-serving-cert\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mcr5\" (UniqueName: \"kubernetes.io/projected/30e6c809-5f33-4c49-bc30-9afda1327eea-kube-api-access-4mcr5\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-serving-cert\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366365 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-client-ca\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc6vt\" (UniqueName: \"kubernetes.io/projected/34bc9da3-b485-489d-a76a-319262f87a8a-kube-api-access-sc6vt\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366429 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366449 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366466 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-csi-data-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366484 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/30e6c809-5f33-4c49-bc30-9afda1327eea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366504 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366520 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhhkx\" (UniqueName: \"kubernetes.io/projected/e899bcbb-0ffd-4989-bba4-47b61fe07832-kube-api-access-nhhkx\") pod \"control-plane-machine-set-operator-78cbb6b69f-njrj9\" (UID: \"e899bcbb-0ffd-4989-bba4-47b61fe07832\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366540 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn745\" (UniqueName: \"kubernetes.io/projected/1262f8ec-a2fc-4b25-b15e-447c35472b6d-kube-api-access-hn745\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366562 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhzp\" (UniqueName: \"kubernetes.io/projected/774d8fe8-cc03-414c-a953-2275e5fd5123-kube-api-access-dkhzp\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366596 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6dbcd2c-e683-4318-860b-447c2df78fdc-config-volume\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366618 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-config\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366641 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366678 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xhcj\" (UniqueName: \"kubernetes.io/projected/d2838428-72ae-469d-884d-6660de5ce2f1-kube-api-access-2xhcj\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366706 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8cdd7414-21ba-4649-9514-66f2e3688022-audit-dir\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366717 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-audit-policies\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.366785 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-certs\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367545 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfwdc\" (UniqueName: \"kubernetes.io/projected/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-kube-api-access-zfwdc\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367598 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-socket-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367627 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/30e6c809-5f33-4c49-bc30-9afda1327eea-proxy-tls\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367705 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/774d8fe8-cc03-414c-a953-2275e5fd5123-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367747 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.367998 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f749021-69b6-4051-bc44-49d2334f087b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q4dmf\" (UID: \"1f749021-69b6-4051-bc44-49d2334f087b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.368024 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.368048 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.368066 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/774d8fe8-cc03-414c-a953-2275e5fd5123-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.368085 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7pcz\" (UniqueName: \"kubernetes.io/projected/5a424e2c-962d-444a-a5a3-57ef0e6133e9-kube-api-access-m7pcz\") pod \"multus-admission-controller-857f4d67dd-t52r4\" (UID: \"5a424e2c-962d-444a-a5a3-57ef0e6133e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.368374 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8cdd7414-21ba-4649-9514-66f2e3688022-audit-dir\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.368731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.369712 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-client-ca\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370034 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370095 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h28nv\" (UniqueName: \"kubernetes.io/projected/8cdd7414-21ba-4649-9514-66f2e3688022-kube-api-access-h28nv\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-config\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370334 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370511 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370562 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370621 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370706 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-plugins-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.370769 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-config-volume\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.371475 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-config\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.372335 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.374027 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dff8795-f189-46d1-8d8b-aae4dad68c63-serving-cert\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.374315 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-config\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.375169 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.375480 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-serving-cert\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.375503 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1262f8ec-a2fc-4b25-b15e-447c35472b6d-srv-cert\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.375836 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.376277 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1262f8ec-a2fc-4b25-b15e-447c35472b6d-profile-collector-cert\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.376846 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e899bcbb-0ffd-4989-bba4-47b61fe07832-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-njrj9\" (UID: \"e899bcbb-0ffd-4989-bba4-47b61fe07832\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.377786 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/774d8fe8-cc03-414c-a953-2275e5fd5123-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.378175 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.378290 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5a424e2c-962d-444a-a5a3-57ef0e6133e9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t52r4\" (UID: \"5a424e2c-962d-444a-a5a3-57ef0e6133e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.378309 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.378676 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.379278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.379470 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1f749021-69b6-4051-bc44-49d2334f087b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-q4dmf\" (UID: \"1f749021-69b6-4051-bc44-49d2334f087b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.379802 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.379841 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.380157 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.381026 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.401273 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfx96\" (UniqueName: \"kubernetes.io/projected/1f749021-69b6-4051-bc44-49d2334f087b-kube-api-access-xfx96\") pod \"package-server-manager-789f6589d5-q4dmf\" (UID: \"1f749021-69b6-4051-bc44-49d2334f087b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.402826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.422141 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2kkb\" (UniqueName: \"kubernetes.io/projected/0dff8795-f189-46d1-8d8b-aae4dad68c63-kube-api-access-m2kkb\") pod \"controller-manager-879f6c89f-xgp2c\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.446495 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfwdc\" (UniqueName: \"kubernetes.io/projected/8fa470e2-2e3d-4fcc-94a5-990b218b6ea3-kube-api-access-zfwdc\") pod \"service-ca-operator-777779d784-swwjr\" (UID: \"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.463206 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd7e0685-11a9-42ba-aedc-e3b88ef9b900-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5s6rc\" (UID: \"dd7e0685-11a9-42ba-aedc-e3b88ef9b900\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.471599 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.471848 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-plugins-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.471883 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-config-volume\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.471929 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pw4c\" (UniqueName: \"kubernetes.io/projected/b6dbcd2c-e683-4318-860b-447c2df78fdc-kube-api-access-8pw4c\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.471955 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4dgs\" (UniqueName: \"kubernetes.io/projected/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-kube-api-access-d4dgs\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.471990 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnw7n\" (UniqueName: \"kubernetes.io/projected/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-kube-api-access-nnw7n\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.472017 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:36.971987281 +0000 UTC m=+138.071993396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472064 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb6t5\" (UniqueName: \"kubernetes.io/projected/34716cea-437d-4d43-86e6-0bdd54bb7ce8-kube-api-access-rb6t5\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472122 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-registration-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472150 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkts2\" (UniqueName: \"kubernetes.io/projected/cc98c09a-446c-45ab-9f91-2f558c153059-kube-api-access-rkts2\") pod \"ingress-canary-bmrff\" (UID: \"cc98c09a-446c-45ab-9f91-2f558c153059\") " pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472178 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc98c09a-446c-45ab-9f91-2f558c153059-cert\") pod \"ingress-canary-bmrff\" (UID: \"cc98c09a-446c-45ab-9f91-2f558c153059\") " pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/34716cea-437d-4d43-86e6-0bdd54bb7ce8-signing-cabundle\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/34716cea-437d-4d43-86e6-0bdd54bb7ce8-signing-key\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472251 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-node-bootstrap-token\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472270 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-mountpoint-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472296 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6dbcd2c-e683-4318-860b-447c2df78fdc-secret-volume\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472322 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-metrics-tls\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472366 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mcr5\" (UniqueName: \"kubernetes.io/projected/30e6c809-5f33-4c49-bc30-9afda1327eea-kube-api-access-4mcr5\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472403 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc6vt\" (UniqueName: \"kubernetes.io/projected/34bc9da3-b485-489d-a76a-319262f87a8a-kube-api-access-sc6vt\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472439 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-csi-data-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472461 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/30e6c809-5f33-4c49-bc30-9afda1327eea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472511 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6dbcd2c-e683-4318-860b-447c2df78fdc-config-volume\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472562 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-certs\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472591 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-socket-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472612 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/30e6c809-5f33-4c49-bc30-9afda1327eea-proxy-tls\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-config-volume\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.472622 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-plugins-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.473510 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-csi-data-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.476475 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6dbcd2c-e683-4318-860b-447c2df78fdc-secret-volume\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.477014 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-registration-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.477295 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/34716cea-437d-4d43-86e6-0bdd54bb7ce8-signing-cabundle\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.477363 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-mountpoint-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.477512 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-metrics-tls\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.477633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6dbcd2c-e683-4318-860b-447c2df78fdc-config-volume\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.478357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/30e6c809-5f33-4c49-bc30-9afda1327eea-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.478427 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/34bc9da3-b485-489d-a76a-319262f87a8a-socket-dir\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.478603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc98c09a-446c-45ab-9f91-2f558c153059-cert\") pod \"ingress-canary-bmrff\" (UID: \"cc98c09a-446c-45ab-9f91-2f558c153059\") " pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.481147 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/34716cea-437d-4d43-86e6-0bdd54bb7ce8-signing-key\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.482877 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-certs\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.483685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/30e6c809-5f33-4c49-bc30-9afda1327eea-proxy-tls\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.486628 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhzp\" (UniqueName: \"kubernetes.io/projected/774d8fe8-cc03-414c-a953-2275e5fd5123-kube-api-access-dkhzp\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.486799 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-node-bootstrap-token\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.509790 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/774d8fe8-cc03-414c-a953-2275e5fd5123-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ztk4c\" (UID: \"774d8fe8-cc03-414c-a953-2275e5fd5123\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.522906 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7pcz\" (UniqueName: \"kubernetes.io/projected/5a424e2c-962d-444a-a5a3-57ef0e6133e9-kube-api-access-m7pcz\") pod \"multus-admission-controller-857f4d67dd-t52r4\" (UID: \"5a424e2c-962d-444a-a5a3-57ef0e6133e9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.544540 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhhkx\" (UniqueName: \"kubernetes.io/projected/e899bcbb-0ffd-4989-bba4-47b61fe07832-kube-api-access-nhhkx\") pod \"control-plane-machine-set-operator-78cbb6b69f-njrj9\" (UID: \"e899bcbb-0ffd-4989-bba4-47b61fe07832\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.569832 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn745\" (UniqueName: \"kubernetes.io/projected/1262f8ec-a2fc-4b25-b15e-447c35472b6d-kube-api-access-hn745\") pod \"catalog-operator-68c6474976-2f58m\" (UID: \"1262f8ec-a2fc-4b25-b15e-447c35472b6d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.574234 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.574507 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.074495719 +0000 UTC m=+138.174501834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.608530 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xhcj\" (UniqueName: \"kubernetes.io/projected/d2838428-72ae-469d-884d-6660de5ce2f1-kube-api-access-2xhcj\") pod \"marketplace-operator-79b997595-25w2m\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.611582 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.618628 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.624241 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h28nv\" (UniqueName: \"kubernetes.io/projected/8cdd7414-21ba-4649-9514-66f2e3688022-kube-api-access-h28nv\") pod \"oauth-openshift-558db77b4-xlkww\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.624991 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.644140 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnw7n\" (UniqueName: \"kubernetes.io/projected/f5aa8a0a-0018-44cc-9f02-ffd0922be33c-kube-api-access-nnw7n\") pod \"dns-default-tmk6k\" (UID: \"f5aa8a0a-0018-44cc-9f02-ffd0922be33c\") " pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.671657 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pw4c\" (UniqueName: \"kubernetes.io/projected/b6dbcd2c-e683-4318-860b-447c2df78fdc-kube-api-access-8pw4c\") pod \"collect-profiles-29420085-47jth\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.672823 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.678636 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.678978 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.178949823 +0000 UTC m=+138.278955938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.679321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.679676 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.179659557 +0000 UTC m=+138.279665672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.687397 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4dgs\" (UniqueName: \"kubernetes.io/projected/b21fbbce-49cb-496d-ba4f-0bae0a5355e9-kube-api-access-d4dgs\") pod \"machine-config-server-297hf\" (UID: \"b21fbbce-49cb-496d-ba4f-0bae0a5355e9\") " pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.688221 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.709439 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkts2\" (UniqueName: \"kubernetes.io/projected/cc98c09a-446c-45ab-9f91-2f558c153059-kube-api-access-rkts2\") pod \"ingress-canary-bmrff\" (UID: \"cc98c09a-446c-45ab-9f91-2f558c153059\") " pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.710100 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.726066 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc6vt\" (UniqueName: \"kubernetes.io/projected/34bc9da3-b485-489d-a76a-319262f87a8a-kube-api-access-sc6vt\") pod \"csi-hostpathplugin-rbk4g\" (UID: \"34bc9da3-b485-489d-a76a-319262f87a8a\") " pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.730067 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.737680 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.747750 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.757960 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.763716 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb6t5\" (UniqueName: \"kubernetes.io/projected/34716cea-437d-4d43-86e6-0bdd54bb7ce8-kube-api-access-rb6t5\") pod \"service-ca-9c57cc56f-h6mkt\" (UID: \"34716cea-437d-4d43-86e6-0bdd54bb7ce8\") " pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.783280 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.783841 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.28380051 +0000 UTC m=+138.383806625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.792021 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.798206 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mcr5\" (UniqueName: \"kubernetes.io/projected/30e6c809-5f33-4c49-bc30-9afda1327eea-kube-api-access-4mcr5\") pod \"machine-config-controller-84d6567774-fh5s4\" (UID: \"30e6c809-5f33-4c49-bc30-9afda1327eea\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.799574 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bmrff" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.810039 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.815279 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-297hf" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.877327 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w5l67" event={"ID":"37ec9532-4218-4e18-bdb2-d9e8781f7cb7","Type":"ContainerStarted","Data":"bbbae67e2052ea973b662b305e40c880c7b57e2d595539cbfdfb3133a3b5142e"} Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.877376 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-w5l67" event={"ID":"37ec9532-4218-4e18-bdb2-d9e8781f7cb7","Type":"ContainerStarted","Data":"970798f2e667e4220c2b2b8a1900ddae671fb159e5412ab4a805513209f13cfd"} Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.884409 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" event={"ID":"f23bacbf-0938-4ecc-bbce-b816bad0a9e0","Type":"ContainerStarted","Data":"464338869e4b7b0884866582b3d54c102fb962946be74164f3b64738fe0156c5"} Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.884539 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.884981 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.384963123 +0000 UTC m=+138.484969238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.904995 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:36 crc kubenswrapper[4894]: I1208 14:48:36.985459 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:36 crc kubenswrapper[4894]: E1208 14:48:36.986897 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.486881192 +0000 UTC m=+138.586887307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.059921 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.067996 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.087186 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.087489 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.587472055 +0000 UTC m=+138.687478170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.190954 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.191391 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.69135203 +0000 UTC m=+138.791358145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.192497 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.193699 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.69314678 +0000 UTC m=+138.793152895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.293953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.294156 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.794126007 +0000 UTC m=+138.894132122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.294209 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.294606 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.794595953 +0000 UTC m=+138.894602138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.382139 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.393147 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:37 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:37 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:37 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.393206 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.396465 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.396858 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.896834712 +0000 UTC m=+138.996840827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.448015 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.498589 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.499109 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:37.999093572 +0000 UTC m=+139.099099687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.508112 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-q6l7x"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.521893 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-77css"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.521934 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.532482 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.537051 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.544918 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.565996 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.572380 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-99z5g"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.574506 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9r5rc"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.579893 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.581275 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w"] Dec 08 14:48:37 crc kubenswrapper[4894]: W1208 14:48:37.588217 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2bb84da_9788_42b9_8c65_900ce3b2b7aa.slice/crio-feeede9fb264bd172d6e56ba9bd6a0884c2eeb79bea695fb061a590279ab6256 WatchSource:0}: Error finding container feeede9fb264bd172d6e56ba9bd6a0884c2eeb79bea695fb061a590279ab6256: Status 404 returned error can't find the container with id feeede9fb264bd172d6e56ba9bd6a0884c2eeb79bea695fb061a590279ab6256 Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.590004 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-955zh"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.597179 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.599681 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.600073 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.100051668 +0000 UTC m=+139.200057783 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.600105 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.600483 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.100471603 +0000 UTC m=+139.200477718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.600989 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-w5l67" podStartSLOduration=119.600978699 podStartE2EDuration="1m59.600978699s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:37.596355323 +0000 UTC m=+138.696361448" watchObservedRunningTime="2025-12-08 14:48:37.600978699 +0000 UTC m=+138.700984814" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.704919 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.705099 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.205069311 +0000 UTC m=+139.305075426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.705559 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.705993 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.205977272 +0000 UTC m=+139.305983387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.736977 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-4scj2"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.747319 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.761898 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7nbxt"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.773727 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.792564 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.794575 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.798796 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-sq5fn"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.800986 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.807752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.808241 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.308220361 +0000 UTC m=+139.408226476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.843943 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9"] Dec 08 14:48:37 crc kubenswrapper[4894]: W1208 14:48:37.879348 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod774d8fe8_cc03_414c_a953_2275e5fd5123.slice/crio-c51aadbeee0fd024b9195c7d85636a35a85e107e1d4e610a25d26875945fc351 WatchSource:0}: Error finding container c51aadbeee0fd024b9195c7d85636a35a85e107e1d4e610a25d26875945fc351: Status 404 returned error can't find the container with id c51aadbeee0fd024b9195c7d85636a35a85e107e1d4e610a25d26875945fc351 Dec 08 14:48:37 crc kubenswrapper[4894]: W1208 14:48:37.882831 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffc6f048_b577_48b6_829d_dee383f4aeb9.slice/crio-210c44db350eb8045426bf48e5ab60bb99b4fe532a66af65aaf9df5537e61c94 WatchSource:0}: Error finding container 210c44db350eb8045426bf48e5ab60bb99b4fe532a66af65aaf9df5537e61c94: Status 404 returned error can't find the container with id 210c44db350eb8045426bf48e5ab60bb99b4fe532a66af65aaf9df5537e61c94 Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.897718 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" event={"ID":"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e","Type":"ContainerStarted","Data":"c427a747d44fb901625d1207e72c57105af0890118ea0779246613d710524bdb"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.897767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" event={"ID":"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e","Type":"ContainerStarted","Data":"1e96bbd862c75e7abfad10c35141628bd6c3cad6d4c3013b9002d517a4e4544d"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.900059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" event={"ID":"f2bb84da-9788-42b9-8c65-900ce3b2b7aa","Type":"ContainerStarted","Data":"feeede9fb264bd172d6e56ba9bd6a0884c2eeb79bea695fb061a590279ab6256"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.905688 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" event={"ID":"0872ad57-8232-471e-a68b-d7550c070794","Type":"ContainerStarted","Data":"66d19ff6700a5e4e8a61f5aee24f78cdf413cf2b6bde0fbcd2337b4eb465c116"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.905968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf"] Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.908714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:37 crc kubenswrapper[4894]: E1208 14:48:37.909036 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.409025142 +0000 UTC m=+139.509031257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:37 crc kubenswrapper[4894]: W1208 14:48:37.910157 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22a82f00_b781_4068_b398_0b4c738db413.slice/crio-959eaca4e12276eac7505f0ac578a964c3a412c8e82539fa67b7d1b4185e9cb2 WatchSource:0}: Error finding container 959eaca4e12276eac7505f0ac578a964c3a412c8e82539fa67b7d1b4185e9cb2: Status 404 returned error can't find the container with id 959eaca4e12276eac7505f0ac578a964c3a412c8e82539fa67b7d1b4185e9cb2 Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.910625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-77css" event={"ID":"368e61a3-8284-46f3-8901-cdcad7a729cb","Type":"ContainerStarted","Data":"a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.910670 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-77css" event={"ID":"368e61a3-8284-46f3-8901-cdcad7a729cb","Type":"ContainerStarted","Data":"48d9156706719c43dbfc8101ade6556d42879d64a082646b371f10bfedf625f3"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.914859 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" event={"ID":"8a9359ef-713b-449b-a468-78b54ccf6c64","Type":"ContainerStarted","Data":"552babc8a050adbea50767c68dda109f97bfff0bb7ff86a476950ae18da64097"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.916440 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" event={"ID":"4231690c-869c-4813-abc1-4aca616a07c6","Type":"ContainerStarted","Data":"4349c2c7ecb20280c05116c23fa53061d67abe3f1ac094f38b69f4557d7942b8"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.921443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" event={"ID":"68be7440-39c0-4161-b248-7bb4f78fa41b","Type":"ContainerStarted","Data":"657e14c27b4337badc0ba26d073c7f57b4b1fe4beb904693df427595f5f34587"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.929563 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-77css" podStartSLOduration=119.929507055 podStartE2EDuration="1m59.929507055s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:37.929035519 +0000 UTC m=+139.029041634" watchObservedRunningTime="2025-12-08 14:48:37.929507055 +0000 UTC m=+139.029513170" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.937251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" event={"ID":"ffc6f048-b577-48b6-829d-dee383f4aeb9","Type":"ContainerStarted","Data":"210c44db350eb8045426bf48e5ab60bb99b4fe532a66af65aaf9df5537e61c94"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.938639 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" event={"ID":"32954141-28e6-48a5-9c87-3d0923fa1afe","Type":"ContainerStarted","Data":"54c764e46e125b828202c98be52790eee85379d91a5741d69a0db27bf709d1a8"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.947036 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-297hf" event={"ID":"b21fbbce-49cb-496d-ba4f-0bae0a5355e9","Type":"ContainerStarted","Data":"43f1e4a7e3f84a3850134364ff5179249e39c560161b6d2446f1521a7232d0b4"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.947073 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-297hf" event={"ID":"b21fbbce-49cb-496d-ba4f-0bae0a5355e9","Type":"ContainerStarted","Data":"a0aa38a7c58103909feb0216c9b323028381f42cabbcaba8523e5745ef8e5485"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.956565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-955zh" event={"ID":"de736b74-9e9d-4ede-acfc-28d64582b060","Type":"ContainerStarted","Data":"98a67d080015e0acd2d5a80cd387bcb12bbbc9552842f824a4e041c252c70695"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.956611 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-955zh" event={"ID":"de736b74-9e9d-4ede-acfc-28d64582b060","Type":"ContainerStarted","Data":"8818374304e3064b3db1ea2c08dfbe2822ae240a0bba45c8a6cad8f4ee21cf40"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.957644 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.959501 4894 patch_prober.go:28] interesting pod/console-operator-58897d9998-955zh container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.959541 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-955zh" podUID="de736b74-9e9d-4ede-acfc-28d64582b060" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.963055 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-297hf" podStartSLOduration=4.96304111 podStartE2EDuration="4.96304111s" podCreationTimestamp="2025-12-08 14:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:37.962962447 +0000 UTC m=+139.062968562" watchObservedRunningTime="2025-12-08 14:48:37.96304111 +0000 UTC m=+139.063047225" Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.963749 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" event={"ID":"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4","Type":"ContainerStarted","Data":"86b79a145dcb944dfaac6c559e5c52cd28e74270283cca7826805d0f8e17ec93"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.964738 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r5rc" event={"ID":"4e9041d3-396d-4331-8b21-82fbf8d7b118","Type":"ContainerStarted","Data":"6f97669d7a19d47041b7f33449d0a21741b89585ca9ed43eed06f764e5a06e07"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.969941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" event={"ID":"03c75f78-3a86-4b5e-8ac1-2a22f80da570","Type":"ContainerStarted","Data":"3776f40ec07d620f3bada15ec3221b98c0e6babc1f39f604bff34822a07bdbc8"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.979575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" event={"ID":"e899bcbb-0ffd-4989-bba4-47b61fe07832","Type":"ContainerStarted","Data":"71d5fd17afde6b5660722ebea395c15266b8b15f85e47c5833e8794b2c15b8aa"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.981121 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" event={"ID":"585fadc5-eeab-4236-b79e-9873939be9e0","Type":"ContainerStarted","Data":"e06f26586a4007bfd4e71435584a8b745e7a9c2fab74988d6beec3a9d38aed2e"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.982488 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" event={"ID":"e6badbf9-7e6b-4c55-a56a-a7153faae4e1","Type":"ContainerStarted","Data":"e8b04840e06eecaf4b5a071196f7357e40f58045d206546a23f7e332280f544a"} Dec 08 14:48:37 crc kubenswrapper[4894]: I1208 14:48:37.993917 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" event={"ID":"64ff7573-1aca-4e17-91f6-ec4f44c86505","Type":"ContainerStarted","Data":"d63b8e1f7ac79de8913cdb64db19ec07ec8f69cfe774476ecaeed2d05713ee92"} Dec 08 14:48:37 crc kubenswrapper[4894]: W1208 14:48:37.998463 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f749021_69b6_4051_bc44_49d2334f087b.slice/crio-bca7b1745c8e7a4543b0fc27c9c1e864b59996786ea3d7f0df3bfa3fd7b6a208 WatchSource:0}: Error finding container bca7b1745c8e7a4543b0fc27c9c1e864b59996786ea3d7f0df3bfa3fd7b6a208: Status 404 returned error can't find the container with id bca7b1745c8e7a4543b0fc27c9c1e864b59996786ea3d7f0df3bfa3fd7b6a208 Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:37.999682 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" event={"ID":"cbb797fc-bf27-449b-b0c2-d72d29fc7845","Type":"ContainerStarted","Data":"bdaff28864554918b621ad1649ecc945ff1df0c20efc1c7fa461900ebbba27f8"} Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.009210 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.010730 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.510706432 +0000 UTC m=+139.610712547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.010881 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" event={"ID":"f23bacbf-0938-4ecc-bbce-b816bad0a9e0","Type":"ContainerStarted","Data":"889a62c761354b4270a4f6c5739b223beae5b4c6ac47c2b66491d54c075d4975"} Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.029317 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-955zh" podStartSLOduration=120.029298341 podStartE2EDuration="2m0.029298341s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:37.991373168 +0000 UTC m=+139.091379283" watchObservedRunningTime="2025-12-08 14:48:38.029298341 +0000 UTC m=+139.129304446" Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.031599 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-f9827" podStartSLOduration=121.031590039 podStartE2EDuration="2m1.031590039s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:38.029221898 +0000 UTC m=+139.129228013" watchObservedRunningTime="2025-12-08 14:48:38.031590039 +0000 UTC m=+139.131596174" Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.071634 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xgp2c"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.087622 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t52r4"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.109973 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-swwjr"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.110488 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.111630 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.611609026 +0000 UTC m=+139.711615201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.128843 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.193742 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tmk6k"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.208214 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25w2m"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.211871 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.212075 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.712034224 +0000 UTC m=+139.812040339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.212135 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.212647 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.712636335 +0000 UTC m=+139.812642640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: W1208 14:48:38.219155 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fa470e2_2e3d_4fcc_94a5_990b218b6ea3.slice/crio-01a26c04b824621983650926a4a6151769b3fe7077e27632d31d202de1910008 WatchSource:0}: Error finding container 01a26c04b824621983650926a4a6151769b3fe7077e27632d31d202de1910008: Status 404 returned error can't find the container with id 01a26c04b824621983650926a4a6151769b3fe7077e27632d31d202de1910008 Dec 08 14:48:38 crc kubenswrapper[4894]: W1208 14:48:38.229441 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d36c61e_d514_4b44_a246_dd24d56c47dd.slice/crio-623bb3ae98597da0723994f9013c5134103c87367e95cf65f295fda2c1c44745 WatchSource:0}: Error finding container 623bb3ae98597da0723994f9013c5134103c87367e95cf65f295fda2c1c44745: Status 404 returned error can't find the container with id 623bb3ae98597da0723994f9013c5134103c87367e95cf65f295fda2c1c44745 Dec 08 14:48:38 crc kubenswrapper[4894]: W1208 14:48:38.244167 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5aa8a0a_0018_44cc_9f02_ffd0922be33c.slice/crio-0c8f50ec3efc6d1f63968f64f2396305886dbb8c091ee8366fbf52863597ac20 WatchSource:0}: Error finding container 0c8f50ec3efc6d1f63968f64f2396305886dbb8c091ee8366fbf52863597ac20: Status 404 returned error can't find the container with id 0c8f50ec3efc6d1f63968f64f2396305886dbb8c091ee8366fbf52863597ac20 Dec 08 14:48:38 crc kubenswrapper[4894]: W1208 14:48:38.250757 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2838428_72ae_469d_884d_6660de5ce2f1.slice/crio-cedab464b16bca7ff79c3af6419111293487fdab87289ae8efdb2f450e47f0f1 WatchSource:0}: Error finding container cedab464b16bca7ff79c3af6419111293487fdab87289ae8efdb2f450e47f0f1: Status 404 returned error can't find the container with id cedab464b16bca7ff79c3af6419111293487fdab87289ae8efdb2f450e47f0f1 Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.304397 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.313176 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.313479 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.813444845 +0000 UTC m=+139.913450960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.313545 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.314059 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.814043706 +0000 UTC m=+139.914049821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.329618 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.389193 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:38 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:38 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:38 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.389852 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.410502 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.418418 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.420384 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:38.920343772 +0000 UTC m=+140.020349887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.425969 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.431637 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bmrff"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.444480 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xlkww"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.457905 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.471356 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rbk4g"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.485831 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-h6mkt"] Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.520045 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.520412 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.020399588 +0000 UTC m=+140.120405703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: W1208 14:48:38.580214 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc98c09a_446c_45ab_9f91_2f558c153059.slice/crio-a067712bc9e6a9d0d649a4dbdb5d60ff64d69faa230b0cef214e5ad37b2d39ed WatchSource:0}: Error finding container a067712bc9e6a9d0d649a4dbdb5d60ff64d69faa230b0cef214e5ad37b2d39ed: Status 404 returned error can't find the container with id a067712bc9e6a9d0d649a4dbdb5d60ff64d69faa230b0cef214e5ad37b2d39ed Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.620545 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.620810 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.120793834 +0000 UTC m=+140.220799949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.721714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.722100 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.222084001 +0000 UTC m=+140.322090156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.823603 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.823907 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.323892556 +0000 UTC m=+140.423898671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:38 crc kubenswrapper[4894]: I1208 14:48:38.925030 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:38 crc kubenswrapper[4894]: E1208 14:48:38.925439 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.425427181 +0000 UTC m=+140.525433296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.017181 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" event={"ID":"30e6c809-5f33-4c49-bc30-9afda1327eea","Type":"ContainerStarted","Data":"dd52c945ee47cfd15775657c8134c112bd943cb9df69cb35caac57ec1023f385"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.018481 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" event={"ID":"ffc6f048-b577-48b6-829d-dee383f4aeb9","Type":"ContainerStarted","Data":"6f52695bd521b06ddab177009db96eab795476401eac6cdb5d106eb49301f45b"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.025493 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" event={"ID":"585fadc5-eeab-4236-b79e-9873939be9e0","Type":"ContainerStarted","Data":"6d07f208c44ba77412abb1d02031ad973d80ada484a512dd11cc5f876ae78be5"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.025798 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.026531 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.026891 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.526863123 +0000 UTC m=+140.626869288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.030841 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" event={"ID":"8a9359ef-713b-449b-a468-78b54ccf6c64","Type":"ContainerStarted","Data":"11fa524756eabe4e6fd17a922f97e69edbd8be4199b765c58e931947450ec31f"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.038089 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" event={"ID":"68be7440-39c0-4161-b248-7bb4f78fa41b","Type":"ContainerStarted","Data":"290d3ba259c90b4f0ece3faec5952537c190fea3a39159aeca8f9e25586f4915"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.045121 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" event={"ID":"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3","Type":"ContainerStarted","Data":"01a26c04b824621983650926a4a6151769b3fe7077e27632d31d202de1910008"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.051936 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9r5rc" event={"ID":"4e9041d3-396d-4331-8b21-82fbf8d7b118","Type":"ContainerStarted","Data":"f4acb96f83aa64d9302d063d8ebc816ed0c6c8b5f0699d9b3d8951f779189fd4"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.052020 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9r5rc" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.055982 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5rc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.056039 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5rc" podUID="4e9041d3-396d-4331-8b21-82fbf8d7b118" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.066156 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" event={"ID":"552313d0-57ae-4e50-ae23-d7a568e1afcd","Type":"ContainerStarted","Data":"e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.066202 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" event={"ID":"552313d0-57ae-4e50-ae23-d7a568e1afcd","Type":"ContainerStarted","Data":"83a49cb88d8f69508a9f341995c2d19fc6e5ffe1f2cf0be00bcec43b913630f2"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.067048 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.068039 4894 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-txhpt container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.068067 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" podUID="552313d0-57ae-4e50-ae23-d7a568e1afcd" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.068870 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" event={"ID":"1262f8ec-a2fc-4b25-b15e-447c35472b6d","Type":"ContainerStarted","Data":"fb8261793727f186d5c8daa5e0f55c3844efac12e1caccd3bab52718d19e8605"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.070269 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bmrff" event={"ID":"cc98c09a-446c-45ab-9f91-2f558c153059","Type":"ContainerStarted","Data":"a067712bc9e6a9d0d649a4dbdb5d60ff64d69faa230b0cef214e5ad37b2d39ed"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.072177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" event={"ID":"b6dbcd2c-e683-4318-860b-447c2df78fdc","Type":"ContainerStarted","Data":"f4f6515778360c5b0575723549605a6e44ac67a25690df8f0924d77f3a370b9e"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.076551 4894 generic.go:334] "Generic (PLEG): container finished" podID="03c75f78-3a86-4b5e-8ac1-2a22f80da570" containerID="43839c2b88206a0725e41417b80105b31f537be5aa330a39ed5379d09fa6c1a0" exitCode=0 Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.076591 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" event={"ID":"03c75f78-3a86-4b5e-8ac1-2a22f80da570","Type":"ContainerDied","Data":"43839c2b88206a0725e41417b80105b31f537be5aa330a39ed5379d09fa6c1a0"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.078112 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-p76j7" podStartSLOduration=121.078102577 podStartE2EDuration="2m1.078102577s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.050418231 +0000 UTC m=+140.150424346" watchObservedRunningTime="2025-12-08 14:48:39.078102577 +0000 UTC m=+140.178108692" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.078649 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-75fl8" podStartSLOduration=121.078644456 podStartE2EDuration="2m1.078644456s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.077841578 +0000 UTC m=+140.177847703" watchObservedRunningTime="2025-12-08 14:48:39.078644456 +0000 UTC m=+140.178650571" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.083059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" event={"ID":"0dff8795-f189-46d1-8d8b-aae4dad68c63","Type":"ContainerStarted","Data":"0014edab7552393d4e55a271efbdff9de8220b76d80436f3ec2a550db3a58838"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.084532 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.084597 4894 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xgp2c container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.084621 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" podUID="0dff8795-f189-46d1-8d8b-aae4dad68c63" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.087121 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" event={"ID":"8d36c61e-d514-4b44-a246-dd24d56c47dd","Type":"ContainerStarted","Data":"623bb3ae98597da0723994f9013c5134103c87367e95cf65f295fda2c1c44745"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.088272 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" event={"ID":"22a82f00-b781-4068-b398-0b4c738db413","Type":"ContainerStarted","Data":"959eaca4e12276eac7505f0ac578a964c3a412c8e82539fa67b7d1b4185e9cb2"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.096394 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" event={"ID":"dd7e0685-11a9-42ba-aedc-e3b88ef9b900","Type":"ContainerStarted","Data":"2fac510e49ed5c7547c707e1e3b32f33c687244695384a1eb1157bf9f01ed285"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.103465 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" event={"ID":"e00f1298-2c1d-4cdd-825f-15f8b9809e0d","Type":"ContainerStarted","Data":"30c797f022bb7835c5b14faa4f3e3873e8ba25b9e0ec999bf9ad620a906c96cb"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.103522 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" event={"ID":"e00f1298-2c1d-4cdd-825f-15f8b9809e0d","Type":"ContainerStarted","Data":"56e01b91a34e94382748c643e3b55775ffb10f036717ef60925aca517829fcef"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.114801 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" event={"ID":"f2bb84da-9788-42b9-8c65-900ce3b2b7aa","Type":"ContainerStarted","Data":"869ba40032031d66be588e9e2ea1d64912c746005c05ccf3697fc4b9ca687831"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.126441 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" podStartSLOduration=121.126419761 podStartE2EDuration="2m1.126419761s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.100409972 +0000 UTC m=+140.200416107" watchObservedRunningTime="2025-12-08 14:48:39.126419761 +0000 UTC m=+140.226425876" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.128646 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.128966 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.628954088 +0000 UTC m=+140.728960203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.137192 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" event={"ID":"2f0d52f8-ca4e-4062-9d57-ff54c976d49e","Type":"ContainerStarted","Data":"a452cee03afbd3c48c2d5814466500d00654ddf70d854df533e6292e80a6fbb7"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.137273 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" event={"ID":"2f0d52f8-ca4e-4062-9d57-ff54c976d49e","Type":"ContainerStarted","Data":"84aacd1f4e3653a4e2bbe8061efd8354f518c2aae6862e321671fd7c0fdc5a4d"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.144752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" event={"ID":"64ff7573-1aca-4e17-91f6-ec4f44c86505","Type":"ContainerStarted","Data":"83d9df625e52e33500f0adc5fec31b8becdba70b9d186e56d251ee671f9ed577"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.147473 4894 generic.go:334] "Generic (PLEG): container finished" podID="4231690c-869c-4813-abc1-4aca616a07c6" containerID="0ed8035a72a409e0be98dcc1a7ee406092fe5886275d2b5135ce65041c84c849" exitCode=0 Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.147617 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" event={"ID":"4231690c-869c-4813-abc1-4aca616a07c6","Type":"ContainerDied","Data":"0ed8035a72a409e0be98dcc1a7ee406092fe5886275d2b5135ce65041c84c849"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.153407 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" podStartSLOduration=121.153391914 podStartE2EDuration="2m1.153391914s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.152989761 +0000 UTC m=+140.252995876" watchObservedRunningTime="2025-12-08 14:48:39.153391914 +0000 UTC m=+140.253398029" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.153828 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" podStartSLOduration=121.153808299 podStartE2EDuration="2m1.153808299s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.127674834 +0000 UTC m=+140.227680949" watchObservedRunningTime="2025-12-08 14:48:39.153808299 +0000 UTC m=+140.253814414" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.162725 4894 generic.go:334] "Generic (PLEG): container finished" podID="0872ad57-8232-471e-a68b-d7550c070794" containerID="ce508d99adfb0ef876dd60f16ae8f5254dacb17da188b8d5ce8ab28b34169930" exitCode=0 Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.162898 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" event={"ID":"0872ad57-8232-471e-a68b-d7550c070794","Type":"ContainerDied","Data":"ce508d99adfb0ef876dd60f16ae8f5254dacb17da188b8d5ce8ab28b34169930"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.164399 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" event={"ID":"d2838428-72ae-469d-884d-6660de5ce2f1","Type":"ContainerStarted","Data":"cedab464b16bca7ff79c3af6419111293487fdab87289ae8efdb2f450e47f0f1"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.178716 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9r5rc" podStartSLOduration=121.178698731 podStartE2EDuration="2m1.178698731s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.17691594 +0000 UTC m=+140.276922055" watchObservedRunningTime="2025-12-08 14:48:39.178698731 +0000 UTC m=+140.278704846" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.186704 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" event={"ID":"774d8fe8-cc03-414c-a953-2275e5fd5123","Type":"ContainerStarted","Data":"c51aadbeee0fd024b9195c7d85636a35a85e107e1d4e610a25d26875945fc351"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.233303 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.233424 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.733396851 +0000 UTC m=+140.833402966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.234080 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.260707 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.760689774 +0000 UTC m=+140.860695889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.303464 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" event={"ID":"32954141-28e6-48a5-9c87-3d0923fa1afe","Type":"ContainerStarted","Data":"29a07c9528e3cb47e408f6dbabbce02ef8cda2288d4bd61f66c379f18378fa55"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.314164 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" event={"ID":"34716cea-437d-4d43-86e6-0bdd54bb7ce8","Type":"ContainerStarted","Data":"ecda2131682fe675e9148bac00957c5fb63e1059b486312529b4829cdaf75949"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.338922 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vx8xs" podStartSLOduration=121.338905992 podStartE2EDuration="2m1.338905992s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.337502994 +0000 UTC m=+140.437509109" watchObservedRunningTime="2025-12-08 14:48:39.338905992 +0000 UTC m=+140.438912107" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.340654 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.341017 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tmk6k" event={"ID":"f5aa8a0a-0018-44cc-9f02-ffd0922be33c","Type":"ContainerStarted","Data":"0c8f50ec3efc6d1f63968f64f2396305886dbb8c091ee8366fbf52863597ac20"} Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.342287 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.842271785 +0000 UTC m=+140.942277900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.366391 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" event={"ID":"8cdd7414-21ba-4649-9514-66f2e3688022","Type":"ContainerStarted","Data":"5d394988f9dd91321666eb557307d51d6dead9310806c66824876e64839bc11f"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.384478 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" event={"ID":"5a424e2c-962d-444a-a5a3-57ef0e6133e9","Type":"ContainerStarted","Data":"9c7f26070e92b7dee277dd1139a20a85a0811eef939e9d1e3e6c82838f9ff56c"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.388849 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:39 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:39 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:39 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.388896 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.389989 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-n2zms" podStartSLOduration=122.389956998 podStartE2EDuration="2m2.389956998s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.386075297 +0000 UTC m=+140.486081412" watchObservedRunningTime="2025-12-08 14:48:39.389956998 +0000 UTC m=+140.489963113" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.405520 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" event={"ID":"1f749021-69b6-4051-bc44-49d2334f087b","Type":"ContainerStarted","Data":"bca7b1745c8e7a4543b0fc27c9c1e864b59996786ea3d7f0df3bfa3fd7b6a208"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.411097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" event={"ID":"34bc9da3-b485-489d-a76a-319262f87a8a","Type":"ContainerStarted","Data":"ebf01046d2a525d96f3bb19ed753a558f0c6e8852f9116ef581b53802264aee4"} Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.444312 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.444914 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:39.944900428 +0000 UTC m=+141.044906543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.456985 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-td4j4" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.465941 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-sq5fn" podStartSLOduration=122.465920488 podStartE2EDuration="2m2.465920488s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:39.464520471 +0000 UTC m=+140.564526587" watchObservedRunningTime="2025-12-08 14:48:39.465920488 +0000 UTC m=+140.565926603" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.504499 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-955zh" Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.551616 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.552194 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.052179537 +0000 UTC m=+141.152185642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.660199 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.660618 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.160600336 +0000 UTC m=+141.260606451 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.762960 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.763633 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.263617361 +0000 UTC m=+141.363623476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:39 crc kubenswrapper[4894]: I1208 14:48:39.864484 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:39 crc kubenswrapper[4894]: E1208 14:48:39.864784 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.364770224 +0000 UTC m=+141.464776339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:39.995341 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:39.996167 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.496151089 +0000 UTC m=+141.596157204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.111152 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.112274 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.612260597 +0000 UTC m=+141.712266712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.215869 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.216725 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.716023078 +0000 UTC m=+141.816029193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.216771 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.217251 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.717238209 +0000 UTC m=+141.817244324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.318572 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.319234 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.819199429 +0000 UTC m=+141.919205554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.398238 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:40 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:40 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:40 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.398563 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.421290 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.421702 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:40.921686477 +0000 UTC m=+142.021692592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.474485 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" event={"ID":"8a9359ef-713b-449b-a468-78b54ccf6c64","Type":"ContainerStarted","Data":"c98117a83a695545c89b039c5fc278e68b67f21aa68d75eaffe918c9d7e53abe"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.488754 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bmrff" event={"ID":"cc98c09a-446c-45ab-9f91-2f558c153059","Type":"ContainerStarted","Data":"4f0729d658dd2c854610bb9080f03587bd4af07d645a446f09dbe6ef2ade3626"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.506583 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9z9w9" podStartSLOduration=122.506565468 podStartE2EDuration="2m2.506565468s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.506077832 +0000 UTC m=+141.606083967" watchObservedRunningTime="2025-12-08 14:48:40.506565468 +0000 UTC m=+141.606571583" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.511009 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" event={"ID":"34bc9da3-b485-489d-a76a-319262f87a8a","Type":"ContainerStarted","Data":"879b2e1c8216f8dc5c1243ac4dd59655945a1691921efbe4ba36fbe4cb09488e"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.516076 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" event={"ID":"65ac3b9e-19fd-4ae8-8c05-bb566ca43be4","Type":"ContainerStarted","Data":"cc0dab56d6c19c8d23749ee478d0b42f4b75863df830bd6c90fd7d5c3800b94a"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.526313 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.526646 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.026629987 +0000 UTC m=+142.126636102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.533516 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" event={"ID":"5a424e2c-962d-444a-a5a3-57ef0e6133e9","Type":"ContainerStarted","Data":"6250ae24da7394c66081cae8aa0b0cc7bf780f07b49ea64bc0370895222642e7"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.542187 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" event={"ID":"8d36c61e-d514-4b44-a246-dd24d56c47dd","Type":"ContainerStarted","Data":"cf757ce99478943f7d0238019c15477ba5f8e86f9c234c057fc6386d618b0d09"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.542238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" event={"ID":"8d36c61e-d514-4b44-a246-dd24d56c47dd","Type":"ContainerStarted","Data":"08cad1bea458213587bcce31bcded70f13c338666051abb61dea312e57579d28"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.550860 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-bmrff" podStartSLOduration=7.550792755 podStartE2EDuration="7.550792755s" podCreationTimestamp="2025-12-08 14:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.53969299 +0000 UTC m=+141.639699105" watchObservedRunningTime="2025-12-08 14:48:40.550792755 +0000 UTC m=+141.650798870" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.553438 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" event={"ID":"30e6c809-5f33-4c49-bc30-9afda1327eea","Type":"ContainerStarted","Data":"b564464823f6f1b1ed05cb9a455b6fa5b33f781902081a4a27d0a9e5b2997808"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.606488 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" event={"ID":"1f749021-69b6-4051-bc44-49d2334f087b","Type":"ContainerStarted","Data":"2a5dbe21028d4de382a17bfcb9206c7c8385b523a969331c0f720eca152526b6"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.606529 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" event={"ID":"1f749021-69b6-4051-bc44-49d2334f087b","Type":"ContainerStarted","Data":"50fb9866f8ecd2a7c7f8257441b8c527779dddcd790c0a6cbfb18f16a322e92b"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.609150 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.637988 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-4scj2" podStartSLOduration=122.637968925 podStartE2EDuration="2m2.637968925s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.613476486 +0000 UTC m=+141.713482601" watchObservedRunningTime="2025-12-08 14:48:40.637968925 +0000 UTC m=+141.737975040" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.638021 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.639924 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.13991222 +0000 UTC m=+142.239918325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.660143 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" event={"ID":"e6badbf9-7e6b-4c55-a56a-a7153faae4e1","Type":"ContainerStarted","Data":"6642a0c40077d97d454a2c5519f741fc3e18e09318cc0b72e38604287e23268e"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.661227 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.672367 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" podStartSLOduration=122.672345957 podStartE2EDuration="2m2.672345957s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.64905482 +0000 UTC m=+141.749060945" watchObservedRunningTime="2025-12-08 14:48:40.672345957 +0000 UTC m=+141.772352072" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.713125 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" podStartSLOduration=122.713111377 podStartE2EDuration="2m2.713111377s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.676723836 +0000 UTC m=+141.776729951" watchObservedRunningTime="2025-12-08 14:48:40.713111377 +0000 UTC m=+141.813117492" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.713619 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h66ct" podStartSLOduration=122.713614454 podStartE2EDuration="2m2.713614454s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.712085442 +0000 UTC m=+141.812091557" watchObservedRunningTime="2025-12-08 14:48:40.713614454 +0000 UTC m=+141.813620569" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.736221 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tmk6k" event={"ID":"f5aa8a0a-0018-44cc-9f02-ffd0922be33c","Type":"ContainerStarted","Data":"4c1b6bde26f1949ed1f4dbec971c247eb1aa6d29e8a5a855cff5971584c20e52"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.736977 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.743255 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.744299 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.244281802 +0000 UTC m=+142.344287917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.763328 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" event={"ID":"8fa470e2-2e3d-4fcc-94a5-990b218b6ea3","Type":"ContainerStarted","Data":"6e4a86ef1f0be6fa320f2489dfa41f8de4baca8dd97b22807f60e2431376d8c9"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.785101 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-tmk6k" podStartSLOduration=7.785081282 podStartE2EDuration="7.785081282s" podCreationTimestamp="2025-12-08 14:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.783096455 +0000 UTC m=+141.883102570" watchObservedRunningTime="2025-12-08 14:48:40.785081282 +0000 UTC m=+141.885087397" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.786302 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" podStartSLOduration=122.786296693 podStartE2EDuration="2m2.786296693s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.75104206 +0000 UTC m=+141.851048175" watchObservedRunningTime="2025-12-08 14:48:40.786296693 +0000 UTC m=+141.886302808" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.792439 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hnsbl" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.800905 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" event={"ID":"774d8fe8-cc03-414c-a953-2275e5fd5123","Type":"ContainerStarted","Data":"2a37ec4a2dc583631685fb78af908cbdfc9999b7c721ca88c04e054a6add28e9"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.813238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" event={"ID":"1262f8ec-a2fc-4b25-b15e-447c35472b6d","Type":"ContainerStarted","Data":"c049f15a44e26ef54795f71745dc92b7a0032bd5a187b4a5aba0017f4644a297"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.814050 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.816168 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" event={"ID":"0dff8795-f189-46d1-8d8b-aae4dad68c63","Type":"ContainerStarted","Data":"653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.822101 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" event={"ID":"64ff7573-1aca-4e17-91f6-ec4f44c86505","Type":"ContainerStarted","Data":"34f57200a1c192a1d665bffe117a057c7d3e261eae0f260f17818b302566774d"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.831244 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.845308 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.846155 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.346139128 +0000 UTC m=+142.446145243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.853575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" event={"ID":"dd7e0685-11a9-42ba-aedc-e3b88ef9b900","Type":"ContainerStarted","Data":"d0b58e528a641a70287958b0ec8033b9a547b9dd3f249a9817a2047f29792863"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.857144 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.867940 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" event={"ID":"4ba8ac2c-7fcb-484e-aae1-b1fa3dc6c08e","Type":"ContainerStarted","Data":"f82ddf4bc5bed8e6b40c8d815041fa6450bfffeb9c86b814de2cda4e05a36b00"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.874349 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-swwjr" podStartSLOduration=122.874329631 podStartE2EDuration="2m2.874329631s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.867352006 +0000 UTC m=+141.967358121" watchObservedRunningTime="2025-12-08 14:48:40.874329631 +0000 UTC m=+141.974335746" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.887551 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" event={"ID":"4231690c-869c-4813-abc1-4aca616a07c6","Type":"ContainerStarted","Data":"c556da10a586f0cd7179959408d786e279eba97ff132e3128998eb6367d6e5ce"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.888468 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.901575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" event={"ID":"34716cea-437d-4d43-86e6-0bdd54bb7ce8","Type":"ContainerStarted","Data":"202d5afc243e244a2c59a9b396292c2649e396e9f08b92b794028cb8d36a4290"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.915916 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-99z5g" podStartSLOduration=122.915896248 podStartE2EDuration="2m2.915896248s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.915333938 +0000 UTC m=+142.015340063" watchObservedRunningTime="2025-12-08 14:48:40.915896248 +0000 UTC m=+142.015902363" Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.934247 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" event={"ID":"22a82f00-b781-4068-b398-0b4c738db413","Type":"ContainerStarted","Data":"8a054d0f9a28945a7f9117994aa1d692c33e861ef48fa2c776e92815179e741a"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.934295 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" event={"ID":"22a82f00-b781-4068-b398-0b4c738db413","Type":"ContainerStarted","Data":"8d4e87b2bab03b1d51e7e60bdebbbed2ba8905af3f1f4efe3f3af5fdfb69e4e3"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.950294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:40 crc kubenswrapper[4894]: E1208 14:48:40.951801 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.451776812 +0000 UTC m=+142.551782977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.978932 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" event={"ID":"cbb797fc-bf27-449b-b0c2-d72d29fc7845","Type":"ContainerStarted","Data":"c636a1906dc656e11907097100adfabd33d971b86be1df9491cf35149ba39f62"} Dec 08 14:48:40 crc kubenswrapper[4894]: I1208 14:48:40.997585 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ztk4c" podStartSLOduration=122.997569761 podStartE2EDuration="2m2.997569761s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:40.996185555 +0000 UTC m=+142.096191680" watchObservedRunningTime="2025-12-08 14:48:40.997569761 +0000 UTC m=+142.097575876" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.000651 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" event={"ID":"d2838428-72ae-469d-884d-6660de5ce2f1","Type":"ContainerStarted","Data":"36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65"} Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.001693 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.006899 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-25w2m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.006983 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" podUID="d2838428-72ae-469d-884d-6660de5ce2f1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.024795 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" event={"ID":"e899bcbb-0ffd-4989-bba4-47b61fe07832","Type":"ContainerStarted","Data":"6669dbfbd13bf47df06ec6b1d3bada14851265fd90c678819ea98146e5f017d1"} Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.052489 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.057187 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" event={"ID":"8cdd7414-21ba-4649-9514-66f2e3688022","Type":"ContainerStarted","Data":"fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf"} Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.058235 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.058386 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.558367618 +0000 UTC m=+142.658373743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.083213 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5s6rc" podStartSLOduration=123.083190878 podStartE2EDuration="2m3.083190878s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.073280883 +0000 UTC m=+142.173286998" watchObservedRunningTime="2025-12-08 14:48:41.083190878 +0000 UTC m=+142.183196993" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.087550 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" event={"ID":"b6dbcd2c-e683-4318-860b-447c2df78fdc","Type":"ContainerStarted","Data":"e396f7b349155beb29b5750cb7eb4eebc6824895cab1545ce41f727124f1ea42"} Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.132446 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" event={"ID":"32954141-28e6-48a5-9c87-3d0923fa1afe","Type":"ContainerStarted","Data":"96bd1ba3d30b7fe566eb0aef0573e2de67d7bb055eb3d078793cf01c27bb9e15"} Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.135972 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5rc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.136029 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5rc" podUID="4e9041d3-396d-4331-8b21-82fbf8d7b118" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.150290 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.153919 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.155238 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.655217886 +0000 UTC m=+142.755224011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.257339 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.271739 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.771710907 +0000 UTC m=+142.871717022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.346706 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2f58m" podStartSLOduration=123.346689943 podStartE2EDuration="2m3.346689943s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.291115683 +0000 UTC m=+142.391121808" watchObservedRunningTime="2025-12-08 14:48:41.346689943 +0000 UTC m=+142.446696058" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.369248 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.369692 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.869676721 +0000 UTC m=+142.969682836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.394035 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:41 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:41 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:41 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.394091 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.446597 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mvkpm" podStartSLOduration=123.446578074 podStartE2EDuration="2m3.446578074s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.382513316 +0000 UTC m=+142.482519431" watchObservedRunningTime="2025-12-08 14:48:41.446578074 +0000 UTC m=+142.546584189" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.476029 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.476337 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:41.97632576 +0000 UTC m=+143.076331875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.521479 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7nbxt" podStartSLOduration=123.521462547 podStartE2EDuration="2m3.521462547s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.520228595 +0000 UTC m=+142.620234720" watchObservedRunningTime="2025-12-08 14:48:41.521462547 +0000 UTC m=+142.621468662" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.521963 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" podStartSLOduration=124.521954744 podStartE2EDuration="2m4.521954744s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.475632756 +0000 UTC m=+142.575638871" watchObservedRunningTime="2025-12-08 14:48:41.521954744 +0000 UTC m=+142.621960869" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.577061 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.577452 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.077435731 +0000 UTC m=+143.177441846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.585059 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" podStartSLOduration=123.585035268 podStartE2EDuration="2m3.585035268s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.551082789 +0000 UTC m=+142.651088914" watchObservedRunningTime="2025-12-08 14:48:41.585035268 +0000 UTC m=+142.685041383" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.585179 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kwjpt" podStartSLOduration=123.585173233 podStartE2EDuration="2m3.585173233s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.574248373 +0000 UTC m=+142.674254498" watchObservedRunningTime="2025-12-08 14:48:41.585173233 +0000 UTC m=+142.685179348" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.675771 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-njrj9" podStartSLOduration=123.675756798 podStartE2EDuration="2m3.675756798s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.618435818 +0000 UTC m=+142.718441933" watchObservedRunningTime="2025-12-08 14:48:41.675756798 +0000 UTC m=+142.775762913" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.676070 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" podStartSLOduration=123.676065378 podStartE2EDuration="2m3.676065378s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.674496025 +0000 UTC m=+142.774502140" watchObservedRunningTime="2025-12-08 14:48:41.676065378 +0000 UTC m=+142.776071493" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.678255 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.678583 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.178571873 +0000 UTC m=+143.278577988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.723452 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-h6mkt" podStartSLOduration=123.723433891 podStartE2EDuration="2m3.723433891s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.722605343 +0000 UTC m=+142.822611468" watchObservedRunningTime="2025-12-08 14:48:41.723433891 +0000 UTC m=+142.823440006" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.768033 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tlvsl" podStartSLOduration=123.768010929 podStartE2EDuration="2m3.768010929s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.767486831 +0000 UTC m=+142.867492956" watchObservedRunningTime="2025-12-08 14:48:41.768010929 +0000 UTC m=+142.868017054" Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.779502 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.780004 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.279985575 +0000 UTC m=+143.379991680 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.881589 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.882029 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.382014616 +0000 UTC m=+143.482020731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:41 crc kubenswrapper[4894]: I1208 14:48:41.982708 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:41 crc kubenswrapper[4894]: E1208 14:48:41.983214 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.483184719 +0000 UTC m=+143.583190844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.051643 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.060791 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" podStartSLOduration=124.060769714 podStartE2EDuration="2m4.060769714s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:41.803610213 +0000 UTC m=+142.903616328" watchObservedRunningTime="2025-12-08 14:48:42.060769714 +0000 UTC m=+143.160775829" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.061653 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xjtc9"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.062737 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.068345 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.087648 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:42 crc kubenswrapper[4894]: E1208 14:48:42.087965 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.587953645 +0000 UTC m=+143.687959760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.090986 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjtc9"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.165699 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t52r4" event={"ID":"5a424e2c-962d-444a-a5a3-57ef0e6133e9","Type":"ContainerStarted","Data":"9e13eb57bc16dad47dea456cb2531eb123463d7dcfaa2294ac120c76afbe479e"} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.172986 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" event={"ID":"30e6c809-5f33-4c49-bc30-9afda1327eea","Type":"ContainerStarted","Data":"bea4b2cc5a8e61bf76772a53938ec5f23f6e65e0b691f8e2628bd73987892d3c"} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.181477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" event={"ID":"34bc9da3-b485-489d-a76a-319262f87a8a","Type":"ContainerStarted","Data":"8ab8ebe7bf27fe8db026a83b68ac4c7fa6975da64fbd7a1791f658166903bc49"} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.191341 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:42 crc kubenswrapper[4894]: E1208 14:48:42.191524 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.691496618 +0000 UTC m=+143.791502733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.191569 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-utilities\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.191619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.191769 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-catalog-content\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.191864 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmcbw\" (UniqueName: \"kubernetes.io/projected/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-kube-api-access-bmcbw\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: E1208 14:48:42.191970 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.691955903 +0000 UTC m=+143.791962048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.192717 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tmk6k" event={"ID":"f5aa8a0a-0018-44cc-9f02-ffd0922be33c","Type":"ContainerStarted","Data":"2161edb084a4e83d5a986661b41d8542105a2061be36e71d52972f77820c6bde"} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.203604 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fh5s4" podStartSLOduration=124.203583926 podStartE2EDuration="2m4.203583926s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:42.200233863 +0000 UTC m=+143.300239978" watchObservedRunningTime="2025-12-08 14:48:42.203583926 +0000 UTC m=+143.303590041" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.216186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" event={"ID":"03c75f78-3a86-4b5e-8ac1-2a22f80da570","Type":"ContainerStarted","Data":"b774494a9a1a54f1e990c16cabb34a938a4798b32bf813ff41ec81a13080f1e9"} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.216236 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" event={"ID":"03c75f78-3a86-4b5e-8ac1-2a22f80da570","Type":"ContainerStarted","Data":"78200ef0e75aff83cd36e4ac6af5617d33cf2f4d7bf7e61eba8d93f6a72aa64f"} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.230128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" event={"ID":"0872ad57-8232-471e-a68b-d7550c070794","Type":"ContainerStarted","Data":"fab2bcaaa10d11f660d6d64f6b58f15e3b05e01372f8c1af9a64818071f3f897"} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.233846 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-25w2m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.233905 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" podUID="d2838428-72ae-469d-884d-6660de5ce2f1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.271638 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t69pb"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.280143 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.286282 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.292753 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.293413 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-catalog-content\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.293569 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmcbw\" (UniqueName: \"kubernetes.io/projected/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-kube-api-access-bmcbw\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.293873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-utilities\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: E1208 14:48:42.295352 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.795322871 +0000 UTC m=+143.895328996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.317198 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-catalog-content\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.317943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.319455 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t69pb"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.319742 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-utilities\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.334877 4894 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 08 14:48:42 crc kubenswrapper[4894]: E1208 14:48:42.351989 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.851972297 +0000 UTC m=+143.951978412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vnpl8" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.375892 4894 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-08T14:48:42.334910929Z","Handler":null,"Name":""} Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.380954 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmcbw\" (UniqueName: \"kubernetes.io/projected/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-kube-api-access-bmcbw\") pod \"certified-operators-xjtc9\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.386419 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:42 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:42 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:42 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.386473 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.390668 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.396161 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" podStartSLOduration=125.396141672 podStartE2EDuration="2m5.396141672s" podCreationTimestamp="2025-12-08 14:46:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:42.320690988 +0000 UTC m=+143.420697103" watchObservedRunningTime="2025-12-08 14:48:42.396141672 +0000 UTC m=+143.496147787" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.424289 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.424932 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-utilities\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.424989 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk84b\" (UniqueName: \"kubernetes.io/projected/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-kube-api-access-wk84b\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.425020 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-catalog-content\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: E1208 14:48:42.425147 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-08 14:48:42.925129122 +0000 UTC m=+144.025135237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.438669 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zntsv"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.439615 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.439702 4894 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.439737 4894 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.445352 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" podStartSLOduration=124.445326616 podStartE2EDuration="2m4.445326616s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:42.43925005 +0000 UTC m=+143.539256165" watchObservedRunningTime="2025-12-08 14:48:42.445326616 +0000 UTC m=+143.545332731" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.492227 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zntsv"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526206 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-utilities\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526258 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526282 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk84b\" (UniqueName: \"kubernetes.io/projected/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-kube-api-access-wk84b\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526303 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-catalog-content\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526346 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66cmn\" (UniqueName: \"kubernetes.io/projected/4efaa068-5a66-4fbc-902d-1d535089a121-kube-api-access-66cmn\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526397 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-utilities\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526435 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-catalog-content\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.526931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-utilities\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.527068 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-catalog-content\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.544675 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.544714 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.582288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk84b\" (UniqueName: \"kubernetes.io/projected/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-kube-api-access-wk84b\") pod \"community-operators-t69pb\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.627391 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-utilities\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.627446 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-catalog-content\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.627521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66cmn\" (UniqueName: \"kubernetes.io/projected/4efaa068-5a66-4fbc-902d-1d535089a121-kube-api-access-66cmn\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.628131 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-utilities\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.628376 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-catalog-content\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.639063 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-65wft"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.639965 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.640946 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.666215 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65wft"] Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.723511 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66cmn\" (UniqueName: \"kubernetes.io/projected/4efaa068-5a66-4fbc-902d-1d535089a121-kube-api-access-66cmn\") pod \"certified-operators-zntsv\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.730925 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-catalog-content\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.730973 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmtnw\" (UniqueName: \"kubernetes.io/projected/07738ca6-502f-43d9-a3bc-951b99d0a0fc-kube-api-access-qmtnw\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.731040 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-utilities\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.757451 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vnpl8\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.792133 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.832039 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.832250 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-utilities\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.832327 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-catalog-content\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.832368 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmtnw\" (UniqueName: \"kubernetes.io/projected/07738ca6-502f-43d9-a3bc-951b99d0a0fc-kube-api-access-qmtnw\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.833158 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-utilities\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.833357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-catalog-content\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.878264 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmtnw\" (UniqueName: \"kubernetes.io/projected/07738ca6-502f-43d9-a3bc-951b99d0a0fc-kube-api-access-qmtnw\") pod \"community-operators-65wft\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.953068 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:42 crc kubenswrapper[4894]: I1208 14:48:42.988091 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.105328 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.107288 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xjtc9"] Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.169344 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.170000 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.199928 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.200173 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.248744 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d37fb499-9aac-4998-a0b6-e545cdca5249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.248852 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d37fb499-9aac-4998-a0b6-e545cdca5249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.250271 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.250744 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.288168 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjtc9" event={"ID":"0edc032d-9fd8-4fbf-b7c7-9106e78404ee","Type":"ContainerStarted","Data":"7ea177201e1e0a3630d94ff8c8d464410613c613aa7cda98e45d8fa2acb84ecd"} Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.330681 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" event={"ID":"34bc9da3-b485-489d-a76a-319262f87a8a","Type":"ContainerStarted","Data":"780d41542623fc70b886ed4272d4fcc5e8b5c96a2dce711e0f0638ec66f6875f"} Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.342727 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t69pb"] Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.350122 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d37fb499-9aac-4998-a0b6-e545cdca5249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.350346 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d37fb499-9aac-4998-a0b6-e545cdca5249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.357848 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.360913 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d37fb499-9aac-4998-a0b6-e545cdca5249-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.386988 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bx4w" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.443035 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d37fb499-9aac-4998-a0b6-e545cdca5249-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.457567 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:43 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:43 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:43 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.457620 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:43 crc kubenswrapper[4894]: I1208 14:48:43.547940 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.076426 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h4hbv"] Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.077964 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.080859 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.101691 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4hbv"] Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.143932 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65wft"] Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.143989 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 08 14:48:44 crc kubenswrapper[4894]: W1208 14:48:44.169973 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07738ca6_502f_43d9_a3bc_951b99d0a0fc.slice/crio-61edb2f22078539a18c1d09a30263584daca09a671e2e48f15f317c1bd8a3bae WatchSource:0}: Error finding container 61edb2f22078539a18c1d09a30263584daca09a671e2e48f15f317c1bd8a3bae: Status 404 returned error can't find the container with id 61edb2f22078539a18c1d09a30263584daca09a671e2e48f15f317c1bd8a3bae Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.175247 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmngr\" (UniqueName: \"kubernetes.io/projected/ccba7943-00e3-422a-91a6-a85439a4c174-kube-api-access-mmngr\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.175306 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-utilities\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.175344 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-catalog-content\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.180406 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vnpl8"] Dec 08 14:48:44 crc kubenswrapper[4894]: W1208 14:48:44.193118 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a9ef065_72ff_4aa4_a729_c4e9884728e6.slice/crio-66db9056d3476b9cb2ceec84a72a0353eb08f2859b16e39578cc7d26f47f707d WatchSource:0}: Error finding container 66db9056d3476b9cb2ceec84a72a0353eb08f2859b16e39578cc7d26f47f707d: Status 404 returned error can't find the container with id 66db9056d3476b9cb2ceec84a72a0353eb08f2859b16e39578cc7d26f47f707d Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.230275 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zntsv"] Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.279281 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-catalog-content\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.279393 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmngr\" (UniqueName: \"kubernetes.io/projected/ccba7943-00e3-422a-91a6-a85439a4c174-kube-api-access-mmngr\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.279431 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-utilities\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.280138 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-catalog-content\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.280286 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-utilities\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.310753 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmngr\" (UniqueName: \"kubernetes.io/projected/ccba7943-00e3-422a-91a6-a85439a4c174-kube-api-access-mmngr\") pod \"redhat-marketplace-h4hbv\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.361066 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t69pb" event={"ID":"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e","Type":"ContainerStarted","Data":"6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.361108 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t69pb" event={"ID":"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e","Type":"ContainerStarted","Data":"913bd1c1fb3b0da6a4cccc41e2cb9e0b2ce96e29d538113c96c734cde7813515"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.362900 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.392202 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:44 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:44 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:44 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.392259 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.396555 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.413554 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" event={"ID":"34bc9da3-b485-489d-a76a-319262f87a8a","Type":"ContainerStarted","Data":"5866a9829629450cac52e051ebd1a78d1fbabbeab64740ecd56e96187ba68cea"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.421405 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zntsv" event={"ID":"4efaa068-5a66-4fbc-902d-1d535089a121","Type":"ContainerStarted","Data":"d0731ab87636cdde1fd0958320b641b093fde856a951256bd10bc3eebde420f9"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.427616 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d37fb499-9aac-4998-a0b6-e545cdca5249","Type":"ContainerStarted","Data":"13dd0feed6a6841c48fa1a9dea99ce087500d4d2a7c7c9e61f0d37e88ed3eb8b"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.428559 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65wft" event={"ID":"07738ca6-502f-43d9-a3bc-951b99d0a0fc","Type":"ContainerStarted","Data":"61edb2f22078539a18c1d09a30263584daca09a671e2e48f15f317c1bd8a3bae"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.433989 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" event={"ID":"8a9ef065-72ff-4aa4-a729-c4e9884728e6","Type":"ContainerStarted","Data":"66db9056d3476b9cb2ceec84a72a0353eb08f2859b16e39578cc7d26f47f707d"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.437266 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jzlrr"] Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.444194 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.453484 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-rbk4g" podStartSLOduration=12.45345321 podStartE2EDuration="12.45345321s" podCreationTimestamp="2025-12-08 14:48:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:44.447438137 +0000 UTC m=+145.547444252" watchObservedRunningTime="2025-12-08 14:48:44.45345321 +0000 UTC m=+145.553459325" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.472561 4894 generic.go:334] "Generic (PLEG): container finished" podID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerID="1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519" exitCode=0 Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.473359 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjtc9" event={"ID":"0edc032d-9fd8-4fbf-b7c7-9106e78404ee","Type":"ContainerDied","Data":"1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519"} Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.501466 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kt7h\" (UniqueName: \"kubernetes.io/projected/17e3e7c4-52dd-48d3-b819-a0fed778803b-kube-api-access-6kt7h\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.501647 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-utilities\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.501703 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-catalog-content\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.511581 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzlrr"] Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.603388 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-utilities\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.603721 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-catalog-content\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.603871 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kt7h\" (UniqueName: \"kubernetes.io/projected/17e3e7c4-52dd-48d3-b819-a0fed778803b-kube-api-access-6kt7h\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.609454 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-utilities\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.621317 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-catalog-content\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.641095 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kt7h\" (UniqueName: \"kubernetes.io/projected/17e3e7c4-52dd-48d3-b819-a0fed778803b-kube-api-access-6kt7h\") pod \"redhat-marketplace-jzlrr\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.687163 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4hbv"] Dec 08 14:48:44 crc kubenswrapper[4894]: I1208 14:48:44.781037 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.167202 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzlrr"] Dec 08 14:48:45 crc kubenswrapper[4894]: W1208 14:48:45.174253 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17e3e7c4_52dd_48d3_b819_a0fed778803b.slice/crio-7b52c1b2ba9e2c592106de50f57531fe5e4faeddb589cf52d528dd37badd604b WatchSource:0}: Error finding container 7b52c1b2ba9e2c592106de50f57531fe5e4faeddb589cf52d528dd37badd604b: Status 404 returned error can't find the container with id 7b52c1b2ba9e2c592106de50f57531fe5e4faeddb589cf52d528dd37badd604b Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.210383 4894 patch_prober.go:28] interesting pod/apiserver-76f77b778f-q6l7x container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]log ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]etcd ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/generic-apiserver-start-informers ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/max-in-flight-filter ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 08 14:48:45 crc kubenswrapper[4894]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/project.openshift.io-projectcache ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-startinformers ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 08 14:48:45 crc kubenswrapper[4894]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 08 14:48:45 crc kubenswrapper[4894]: livez check failed Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.210778 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" podUID="03c75f78-3a86-4b5e-8ac1-2a22f80da570" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.210869 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.210967 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.211790 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.212679 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.212705 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.216759 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.286534 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.286593 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.289431 4894 patch_prober.go:28] interesting pod/console-f9d7485db-77css container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.289481 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-77css" podUID="368e61a3-8284-46f3-8901-cdcad7a729cb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.311916 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.311999 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.313313 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.315721 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.316327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.321187 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.321906 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5rc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.321940 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9r5rc" podUID="4e9041d3-396d-4331-8b21-82fbf8d7b118" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.322015 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-9r5rc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.322096 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9r5rc" podUID="4e9041d3-396d-4331-8b21-82fbf8d7b118" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.385162 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:45 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:45 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:45 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.385224 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.488722 4894 generic.go:334] "Generic (PLEG): container finished" podID="ccba7943-00e3-422a-91a6-a85439a4c174" containerID="5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021" exitCode=0 Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.488861 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4hbv" event={"ID":"ccba7943-00e3-422a-91a6-a85439a4c174","Type":"ContainerDied","Data":"5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.489110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4hbv" event={"ID":"ccba7943-00e3-422a-91a6-a85439a4c174","Type":"ContainerStarted","Data":"d5881c879b060af61c583a68a46d55fac4fc6e188e8fd23414cc621a9dd19bdb"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.491724 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" event={"ID":"8a9ef065-72ff-4aa4-a729-c4e9884728e6","Type":"ContainerStarted","Data":"be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.492134 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.494629 4894 generic.go:334] "Generic (PLEG): container finished" podID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerID="6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04" exitCode=0 Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.494692 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t69pb" event={"ID":"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e","Type":"ContainerDied","Data":"6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.499246 4894 generic.go:334] "Generic (PLEG): container finished" podID="4efaa068-5a66-4fbc-902d-1d535089a121" containerID="7a8bc8bb8edaaeaf0fbf6eeaecdeabd21258334f4416716c4718c7df17a30934" exitCode=0 Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.500416 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zntsv" event={"ID":"4efaa068-5a66-4fbc-902d-1d535089a121","Type":"ContainerDied","Data":"7a8bc8bb8edaaeaf0fbf6eeaecdeabd21258334f4416716c4718c7df17a30934"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.506383 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d37fb499-9aac-4998-a0b6-e545cdca5249","Type":"ContainerStarted","Data":"9bdd0ed3fdab532bf79cbed8188223811dc560d258b5b2f33065467123acb6d0"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.509730 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.515674 4894 generic.go:334] "Generic (PLEG): container finished" podID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerID="b84136b3cd76252e4ce31d0e91a826ba2d9de867aa226518723fd0957d49452e" exitCode=0 Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.515972 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65wft" event={"ID":"07738ca6-502f-43d9-a3bc-951b99d0a0fc","Type":"ContainerDied","Data":"b84136b3cd76252e4ce31d0e91a826ba2d9de867aa226518723fd0957d49452e"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.518419 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzlrr" event={"ID":"17e3e7c4-52dd-48d3-b819-a0fed778803b","Type":"ContainerStarted","Data":"7b52c1b2ba9e2c592106de50f57531fe5e4faeddb589cf52d528dd37badd604b"} Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.524055 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.5240344820000002 podStartE2EDuration="2.524034482s" podCreationTimestamp="2025-12-08 14:48:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:45.522619875 +0000 UTC m=+146.622626010" watchObservedRunningTime="2025-12-08 14:48:45.524034482 +0000 UTC m=+146.624040607" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.581544 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" podStartSLOduration=127.581519638 podStartE2EDuration="2m7.581519638s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:45.573421084 +0000 UTC m=+146.673427199" watchObservedRunningTime="2025-12-08 14:48:45.581519638 +0000 UTC m=+146.681525753" Dec 08 14:48:45 crc kubenswrapper[4894]: W1208 14:48:45.608573 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-34785529e0540285a732d05f6471c8a52f6fdea2389c2fda4e0a575769c9acba WatchSource:0}: Error finding container 34785529e0540285a732d05f6471c8a52f6fdea2389c2fda4e0a575769c9acba: Status 404 returned error can't find the container with id 34785529e0540285a732d05f6471c8a52f6fdea2389c2fda4e0a575769c9acba Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.644216 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gbbpn"] Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.645707 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.648244 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 08 14:48:45 crc kubenswrapper[4894]: W1208 14:48:45.654053 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-3ca4649a7eb01d416592d25dd04daef097a5d886837bd76502b686de93dbfc23 WatchSource:0}: Error finding container 3ca4649a7eb01d416592d25dd04daef097a5d886837bd76502b686de93dbfc23: Status 404 returned error can't find the container with id 3ca4649a7eb01d416592d25dd04daef097a5d886837bd76502b686de93dbfc23 Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.666026 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gbbpn"] Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.724459 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnztk\" (UniqueName: \"kubernetes.io/projected/a4d8a8de-7e5f-4387-8a9f-281421db9641-kube-api-access-tnztk\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.724516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-catalog-content\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.724577 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-utilities\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.829848 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnztk\" (UniqueName: \"kubernetes.io/projected/a4d8a8de-7e5f-4387-8a9f-281421db9641-kube-api-access-tnztk\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.830779 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-catalog-content\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.831290 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-utilities\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.832295 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-catalog-content\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.832363 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-utilities\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.849978 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnztk\" (UniqueName: \"kubernetes.io/projected/a4d8a8de-7e5f-4387-8a9f-281421db9641-kube-api-access-tnztk\") pod \"redhat-operators-gbbpn\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:45 crc kubenswrapper[4894]: I1208 14:48:45.968444 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.031969 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tglg6"] Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.033232 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.046365 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tglg6"] Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.140460 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nbzv\" (UniqueName: \"kubernetes.io/projected/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-kube-api-access-6nbzv\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.140830 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-catalog-content\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.140870 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-utilities\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.153718 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.154406 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.169340 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.207254 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gbbpn"] Dec 08 14:48:46 crc kubenswrapper[4894]: W1208 14:48:46.214831 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4d8a8de_7e5f_4387_8a9f_281421db9641.slice/crio-20e2e6db693382400d817579908639f3a7affd67972008a0470505f6043715f5 WatchSource:0}: Error finding container 20e2e6db693382400d817579908639f3a7affd67972008a0470505f6043715f5: Status 404 returned error can't find the container with id 20e2e6db693382400d817579908639f3a7affd67972008a0470505f6043715f5 Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.242352 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nbzv\" (UniqueName: \"kubernetes.io/projected/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-kube-api-access-6nbzv\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.242542 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-catalog-content\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.242602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-utilities\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.243044 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-utilities\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.244571 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-catalog-content\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.260758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nbzv\" (UniqueName: \"kubernetes.io/projected/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-kube-api-access-6nbzv\") pod \"redhat-operators-tglg6\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.380970 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.383565 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:46 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:46 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:46 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.383627 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.452211 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.537979 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ce0f8c99cbeadedf33e74131a8961baa16ab9b589af33a5d596804899bcf01f0"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.538030 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"34785529e0540285a732d05f6471c8a52f6fdea2389c2fda4e0a575769c9acba"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.538297 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.542612 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4c8b6bd4bb9299f9bcbd0c8fae161acea08d909f6fb454496185ef31cde94ae5"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.542669 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3ca4649a7eb01d416592d25dd04daef097a5d886837bd76502b686de93dbfc23"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.549541 4894 generic.go:334] "Generic (PLEG): container finished" podID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerID="550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413" exitCode=0 Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.550205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzlrr" event={"ID":"17e3e7c4-52dd-48d3-b819-a0fed778803b","Type":"ContainerDied","Data":"550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.561273 4894 generic.go:334] "Generic (PLEG): container finished" podID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerID="dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646" exitCode=0 Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.561355 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbbpn" event={"ID":"a4d8a8de-7e5f-4387-8a9f-281421db9641","Type":"ContainerDied","Data":"dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.561394 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbbpn" event={"ID":"a4d8a8de-7e5f-4387-8a9f-281421db9641","Type":"ContainerStarted","Data":"20e2e6db693382400d817579908639f3a7affd67972008a0470505f6043715f5"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.563659 4894 generic.go:334] "Generic (PLEG): container finished" podID="d37fb499-9aac-4998-a0b6-e545cdca5249" containerID="9bdd0ed3fdab532bf79cbed8188223811dc560d258b5b2f33065467123acb6d0" exitCode=0 Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.563752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d37fb499-9aac-4998-a0b6-e545cdca5249","Type":"ContainerDied","Data":"9bdd0ed3fdab532bf79cbed8188223811dc560d258b5b2f33065467123acb6d0"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.583077 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7d10d3839d7915a9a0bd365f7959e9ece11cda4e2f1601d8a0007dd39576c147"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.583132 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5f261c3b6dc938a2f0937146962ac6b62099d2882da87b2643d5cde6c43bee5c"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.602091 4894 generic.go:334] "Generic (PLEG): container finished" podID="b6dbcd2c-e683-4318-860b-447c2df78fdc" containerID="e396f7b349155beb29b5750cb7eb4eebc6824895cab1545ce41f727124f1ea42" exitCode=0 Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.602446 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" event={"ID":"b6dbcd2c-e683-4318-860b-447c2df78fdc","Type":"ContainerDied","Data":"e396f7b349155beb29b5750cb7eb4eebc6824895cab1545ce41f727124f1ea42"} Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.624023 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l2fzd" Dec 08 14:48:46 crc kubenswrapper[4894]: I1208 14:48:46.791279 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tglg6"] Dec 08 14:48:47 crc kubenswrapper[4894]: I1208 14:48:47.385316 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:47 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:47 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:47 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:47 crc kubenswrapper[4894]: I1208 14:48:47.385378 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:47 crc kubenswrapper[4894]: I1208 14:48:47.616795 4894 generic.go:334] "Generic (PLEG): container finished" podID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerID="d039f2a346e669e9d9bf10cb85802a4847ac842ea2447ab2f2490c71e33ac212" exitCode=0 Dec 08 14:48:47 crc kubenswrapper[4894]: I1208 14:48:47.616961 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tglg6" event={"ID":"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1","Type":"ContainerDied","Data":"d039f2a346e669e9d9bf10cb85802a4847ac842ea2447ab2f2490c71e33ac212"} Dec 08 14:48:47 crc kubenswrapper[4894]: I1208 14:48:47.617025 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tglg6" event={"ID":"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1","Type":"ContainerStarted","Data":"193db2d821d18104791d5b683870e4efac1027ac25948221a7b3aa6bfe654d4b"} Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.041209 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.087471 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.090619 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6dbcd2c-e683-4318-860b-447c2df78fdc-secret-volume\") pod \"b6dbcd2c-e683-4318-860b-447c2df78fdc\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.090688 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pw4c\" (UniqueName: \"kubernetes.io/projected/b6dbcd2c-e683-4318-860b-447c2df78fdc-kube-api-access-8pw4c\") pod \"b6dbcd2c-e683-4318-860b-447c2df78fdc\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.090797 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6dbcd2c-e683-4318-860b-447c2df78fdc-config-volume\") pod \"b6dbcd2c-e683-4318-860b-447c2df78fdc\" (UID: \"b6dbcd2c-e683-4318-860b-447c2df78fdc\") " Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.091726 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6dbcd2c-e683-4318-860b-447c2df78fdc-config-volume" (OuterVolumeSpecName: "config-volume") pod "b6dbcd2c-e683-4318-860b-447c2df78fdc" (UID: "b6dbcd2c-e683-4318-860b-447c2df78fdc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.103284 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6dbcd2c-e683-4318-860b-447c2df78fdc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b6dbcd2c-e683-4318-860b-447c2df78fdc" (UID: "b6dbcd2c-e683-4318-860b-447c2df78fdc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.108363 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6dbcd2c-e683-4318-860b-447c2df78fdc-kube-api-access-8pw4c" (OuterVolumeSpecName: "kube-api-access-8pw4c") pod "b6dbcd2c-e683-4318-860b-447c2df78fdc" (UID: "b6dbcd2c-e683-4318-860b-447c2df78fdc"). InnerVolumeSpecName "kube-api-access-8pw4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.193267 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d37fb499-9aac-4998-a0b6-e545cdca5249-kube-api-access\") pod \"d37fb499-9aac-4998-a0b6-e545cdca5249\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.193624 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d37fb499-9aac-4998-a0b6-e545cdca5249-kubelet-dir\") pod \"d37fb499-9aac-4998-a0b6-e545cdca5249\" (UID: \"d37fb499-9aac-4998-a0b6-e545cdca5249\") " Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.193787 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d37fb499-9aac-4998-a0b6-e545cdca5249-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d37fb499-9aac-4998-a0b6-e545cdca5249" (UID: "d37fb499-9aac-4998-a0b6-e545cdca5249"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.194283 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6dbcd2c-e683-4318-860b-447c2df78fdc-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.194310 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pw4c\" (UniqueName: \"kubernetes.io/projected/b6dbcd2c-e683-4318-860b-447c2df78fdc-kube-api-access-8pw4c\") on node \"crc\" DevicePath \"\"" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.194482 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d37fb499-9aac-4998-a0b6-e545cdca5249-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.194562 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6dbcd2c-e683-4318-860b-447c2df78fdc-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.197061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37fb499-9aac-4998-a0b6-e545cdca5249-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d37fb499-9aac-4998-a0b6-e545cdca5249" (UID: "d37fb499-9aac-4998-a0b6-e545cdca5249"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.296778 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d37fb499-9aac-4998-a0b6-e545cdca5249-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.383644 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:48 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:48 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:48 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.383710 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.652051 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" event={"ID":"b6dbcd2c-e683-4318-860b-447c2df78fdc","Type":"ContainerDied","Data":"f4f6515778360c5b0575723549605a6e44ac67a25690df8f0924d77f3a370b9e"} Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.652126 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4f6515778360c5b0575723549605a6e44ac67a25690df8f0924d77f3a370b9e" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.652081 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.658105 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d37fb499-9aac-4998-a0b6-e545cdca5249","Type":"ContainerDied","Data":"13dd0feed6a6841c48fa1a9dea99ce087500d4d2a7c7c9e61f0d37e88ed3eb8b"} Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.658127 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.658149 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13dd0feed6a6841c48fa1a9dea99ce087500d4d2a7c7c9e61f0d37e88ed3eb8b" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.750233 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 08 14:48:48 crc kubenswrapper[4894]: E1208 14:48:48.750850 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37fb499-9aac-4998-a0b6-e545cdca5249" containerName="pruner" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.750871 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37fb499-9aac-4998-a0b6-e545cdca5249" containerName="pruner" Dec 08 14:48:48 crc kubenswrapper[4894]: E1208 14:48:48.750893 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6dbcd2c-e683-4318-860b-447c2df78fdc" containerName="collect-profiles" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.750899 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6dbcd2c-e683-4318-860b-447c2df78fdc" containerName="collect-profiles" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.751013 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6dbcd2c-e683-4318-860b-447c2df78fdc" containerName="collect-profiles" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.751033 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37fb499-9aac-4998-a0b6-e545cdca5249" containerName="pruner" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.751449 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.754475 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.754664 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.761921 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.802265 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/42ba905c-6dd6-41c1-8703-6160ee507486-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.802323 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42ba905c-6dd6-41c1-8703-6160ee507486-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.904169 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/42ba905c-6dd6-41c1-8703-6160ee507486-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.904298 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/42ba905c-6dd6-41c1-8703-6160ee507486-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.904360 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42ba905c-6dd6-41c1-8703-6160ee507486-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:48 crc kubenswrapper[4894]: I1208 14:48:48.921113 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42ba905c-6dd6-41c1-8703-6160ee507486-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:49 crc kubenswrapper[4894]: I1208 14:48:49.068590 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:49 crc kubenswrapper[4894]: I1208 14:48:49.386256 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:49 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:49 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:49 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:49 crc kubenswrapper[4894]: I1208 14:48:49.386325 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:49 crc kubenswrapper[4894]: I1208 14:48:49.579909 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 08 14:48:49 crc kubenswrapper[4894]: I1208 14:48:49.669411 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"42ba905c-6dd6-41c1-8703-6160ee507486","Type":"ContainerStarted","Data":"11931618b191f96dc78093be51f3ed506d027b5baccb18f8c4b35bfb89b779a4"} Dec 08 14:48:50 crc kubenswrapper[4894]: I1208 14:48:50.206301 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:50 crc kubenswrapper[4894]: I1208 14:48:50.210585 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-q6l7x" Dec 08 14:48:50 crc kubenswrapper[4894]: I1208 14:48:50.384791 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:50 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:50 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:50 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:50 crc kubenswrapper[4894]: I1208 14:48:50.385145 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:50 crc kubenswrapper[4894]: I1208 14:48:50.729644 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"42ba905c-6dd6-41c1-8703-6160ee507486","Type":"ContainerStarted","Data":"4699d28bafec7171f162b84e7edbfd93cc5c4502c9258b76c31458b45dceae4d"} Dec 08 14:48:51 crc kubenswrapper[4894]: I1208 14:48:51.382254 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:51 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:51 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:51 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:51 crc kubenswrapper[4894]: I1208 14:48:51.382319 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:51 crc kubenswrapper[4894]: I1208 14:48:51.792116 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.79209885 podStartE2EDuration="3.79209885s" podCreationTimestamp="2025-12-08 14:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:48:51.791568282 +0000 UTC m=+152.891574417" watchObservedRunningTime="2025-12-08 14:48:51.79209885 +0000 UTC m=+152.892104965" Dec 08 14:48:51 crc kubenswrapper[4894]: I1208 14:48:51.813121 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-tmk6k" Dec 08 14:48:52 crc kubenswrapper[4894]: I1208 14:48:52.383087 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:52 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:52 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:52 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:52 crc kubenswrapper[4894]: I1208 14:48:52.383156 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:52 crc kubenswrapper[4894]: I1208 14:48:52.782545 4894 generic.go:334] "Generic (PLEG): container finished" podID="42ba905c-6dd6-41c1-8703-6160ee507486" containerID="4699d28bafec7171f162b84e7edbfd93cc5c4502c9258b76c31458b45dceae4d" exitCode=0 Dec 08 14:48:52 crc kubenswrapper[4894]: I1208 14:48:52.782594 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"42ba905c-6dd6-41c1-8703-6160ee507486","Type":"ContainerDied","Data":"4699d28bafec7171f162b84e7edbfd93cc5c4502c9258b76c31458b45dceae4d"} Dec 08 14:48:53 crc kubenswrapper[4894]: I1208 14:48:53.382860 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:53 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:53 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:53 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:53 crc kubenswrapper[4894]: I1208 14:48:53.383410 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:54 crc kubenswrapper[4894]: I1208 14:48:54.383714 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:54 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:54 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:54 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:54 crc kubenswrapper[4894]: I1208 14:48:54.383807 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:55 crc kubenswrapper[4894]: I1208 14:48:55.287244 4894 patch_prober.go:28] interesting pod/console-f9d7485db-77css container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Dec 08 14:48:55 crc kubenswrapper[4894]: I1208 14:48:55.287614 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-77css" podUID="368e61a3-8284-46f3-8901-cdcad7a729cb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.12:8443/health\": dial tcp 10.217.0.12:8443: connect: connection refused" Dec 08 14:48:55 crc kubenswrapper[4894]: I1208 14:48:55.328598 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9r5rc" Dec 08 14:48:55 crc kubenswrapper[4894]: I1208 14:48:55.383319 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:55 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:55 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:55 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:55 crc kubenswrapper[4894]: I1208 14:48:55.383423 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:56 crc kubenswrapper[4894]: I1208 14:48:56.384121 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:56 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:56 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:56 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:56 crc kubenswrapper[4894]: I1208 14:48:56.384191 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:57 crc kubenswrapper[4894]: I1208 14:48:57.385559 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:57 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 08 14:48:57 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:57 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:57 crc kubenswrapper[4894]: I1208 14:48:57.385926 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.383270 4894 patch_prober.go:28] interesting pod/router-default-5444994796-w5l67 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 08 14:48:58 crc kubenswrapper[4894]: [+]has-synced ok Dec 08 14:48:58 crc kubenswrapper[4894]: [+]process-running ok Dec 08 14:48:58 crc kubenswrapper[4894]: healthz check failed Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.383738 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-w5l67" podUID="37ec9532-4218-4e18-bdb2-d9e8781f7cb7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.789169 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.821368 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"42ba905c-6dd6-41c1-8703-6160ee507486","Type":"ContainerDied","Data":"11931618b191f96dc78093be51f3ed506d027b5baccb18f8c4b35bfb89b779a4"} Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.821416 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11931618b191f96dc78093be51f3ed506d027b5baccb18f8c4b35bfb89b779a4" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.821479 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.883242 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42ba905c-6dd6-41c1-8703-6160ee507486-kube-api-access\") pod \"42ba905c-6dd6-41c1-8703-6160ee507486\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.883393 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/42ba905c-6dd6-41c1-8703-6160ee507486-kubelet-dir\") pod \"42ba905c-6dd6-41c1-8703-6160ee507486\" (UID: \"42ba905c-6dd6-41c1-8703-6160ee507486\") " Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.883543 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/42ba905c-6dd6-41c1-8703-6160ee507486-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "42ba905c-6dd6-41c1-8703-6160ee507486" (UID: "42ba905c-6dd6-41c1-8703-6160ee507486"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.883802 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/42ba905c-6dd6-41c1-8703-6160ee507486-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.889219 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42ba905c-6dd6-41c1-8703-6160ee507486-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "42ba905c-6dd6-41c1-8703-6160ee507486" (UID: "42ba905c-6dd6-41c1-8703-6160ee507486"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:48:58 crc kubenswrapper[4894]: I1208 14:48:58.985436 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42ba905c-6dd6-41c1-8703-6160ee507486-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:48:59 crc kubenswrapper[4894]: I1208 14:48:59.383406 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:48:59 crc kubenswrapper[4894]: I1208 14:48:59.386919 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-w5l67" Dec 08 14:49:00 crc kubenswrapper[4894]: I1208 14:49:00.304478 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:49:00 crc kubenswrapper[4894]: I1208 14:49:00.310255 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a05fc4-1361-4ebc-891a-e2b49df28ffa-metrics-certs\") pod \"network-metrics-daemon-z2zz6\" (UID: \"49a05fc4-1361-4ebc-891a-e2b49df28ffa\") " pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:49:00 crc kubenswrapper[4894]: I1208 14:49:00.327725 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2zz6" Dec 08 14:49:02 crc kubenswrapper[4894]: I1208 14:49:02.962124 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:49:05 crc kubenswrapper[4894]: I1208 14:49:05.338687 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:49:05 crc kubenswrapper[4894]: I1208 14:49:05.342089 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:49:07 crc kubenswrapper[4894]: I1208 14:49:07.297003 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:49:07 crc kubenswrapper[4894]: I1208 14:49:07.297078 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:49:16 crc kubenswrapper[4894]: I1208 14:49:16.406411 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-q4dmf" Dec 08 14:49:16 crc kubenswrapper[4894]: E1208 14:49:16.558244 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 08 14:49:16 crc kubenswrapper[4894]: E1208 14:49:16.558995 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bmcbw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-xjtc9_openshift-marketplace(0edc032d-9fd8-4fbf-b7c7-9106e78404ee): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 14:49:16 crc kubenswrapper[4894]: E1208 14:49:16.560379 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-xjtc9" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" Dec 08 14:49:16 crc kubenswrapper[4894]: E1208 14:49:16.581663 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 08 14:49:16 crc kubenswrapper[4894]: E1208 14:49:16.581868 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tnztk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gbbpn_openshift-marketplace(a4d8a8de-7e5f-4387-8a9f-281421db9641): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 14:49:16 crc kubenswrapper[4894]: E1208 14:49:16.583966 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gbbpn" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" Dec 08 14:49:17 crc kubenswrapper[4894]: E1208 14:49:17.715200 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-xjtc9" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" Dec 08 14:49:17 crc kubenswrapper[4894]: E1208 14:49:17.715348 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gbbpn" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" Dec 08 14:49:17 crc kubenswrapper[4894]: E1208 14:49:17.784321 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 08 14:49:17 crc kubenswrapper[4894]: E1208 14:49:17.784576 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmtnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-65wft_openshift-marketplace(07738ca6-502f-43d9-a3bc-951b99d0a0fc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 14:49:17 crc kubenswrapper[4894]: E1208 14:49:17.786418 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-65wft" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" Dec 08 14:49:18 crc kubenswrapper[4894]: E1208 14:49:18.947474 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-65wft" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.021040 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.021335 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6kt7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jzlrr_openshift-marketplace(17e3e7c4-52dd-48d3-b819-a0fed778803b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.022462 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jzlrr" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.059666 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.060063 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-66cmn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zntsv_openshift-marketplace(4efaa068-5a66-4fbc-902d-1d535089a121): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.066091 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zntsv" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.071778 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.071960 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wk84b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-t69pb_openshift-marketplace(fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.073130 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-t69pb" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.074803 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.074962 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mmngr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-h4hbv_openshift-marketplace(ccba7943-00e3-422a-91a6-a85439a4c174): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.076468 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-h4hbv" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" Dec 08 14:49:19 crc kubenswrapper[4894]: I1208 14:49:19.395971 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-z2zz6"] Dec 08 14:49:19 crc kubenswrapper[4894]: W1208 14:49:19.398272 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49a05fc4_1361_4ebc_891a_e2b49df28ffa.slice/crio-16582edb6b94b0c44a489de0b1b6aff0f727f2f2fdbf3e1358bd070e5b28825a WatchSource:0}: Error finding container 16582edb6b94b0c44a489de0b1b6aff0f727f2f2fdbf3e1358bd070e5b28825a: Status 404 returned error can't find the container with id 16582edb6b94b0c44a489de0b1b6aff0f727f2f2fdbf3e1358bd070e5b28825a Dec 08 14:49:19 crc kubenswrapper[4894]: I1208 14:49:19.952642 4894 generic.go:334] "Generic (PLEG): container finished" podID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerID="cef07a9e0a10da78165c7115d227f0de203dce4e6433af592118abf805f27d81" exitCode=0 Dec 08 14:49:19 crc kubenswrapper[4894]: I1208 14:49:19.952871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tglg6" event={"ID":"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1","Type":"ContainerDied","Data":"cef07a9e0a10da78165c7115d227f0de203dce4e6433af592118abf805f27d81"} Dec 08 14:49:19 crc kubenswrapper[4894]: I1208 14:49:19.958358 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" event={"ID":"49a05fc4-1361-4ebc-891a-e2b49df28ffa","Type":"ContainerStarted","Data":"a63dd4d95471e104738855893e98dff6359c7bb03c3accfeb9a1d5dd96d3275e"} Dec 08 14:49:19 crc kubenswrapper[4894]: I1208 14:49:19.958724 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" event={"ID":"49a05fc4-1361-4ebc-891a-e2b49df28ffa","Type":"ContainerStarted","Data":"b6805dc553febc048d0dd5f51087bb0f3a812dde82c065a3a2a883d6a48a0b6b"} Dec 08 14:49:19 crc kubenswrapper[4894]: I1208 14:49:19.958745 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z2zz6" event={"ID":"49a05fc4-1361-4ebc-891a-e2b49df28ffa","Type":"ContainerStarted","Data":"16582edb6b94b0c44a489de0b1b6aff0f727f2f2fdbf3e1358bd070e5b28825a"} Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.960801 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-h4hbv" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.960907 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jzlrr" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.960966 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-t69pb" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" Dec 08 14:49:19 crc kubenswrapper[4894]: E1208 14:49:19.961086 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zntsv" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" Dec 08 14:49:20 crc kubenswrapper[4894]: I1208 14:49:20.027183 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-z2zz6" podStartSLOduration=162.027153931 podStartE2EDuration="2m42.027153931s" podCreationTimestamp="2025-12-08 14:46:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:49:20.024736269 +0000 UTC m=+181.124742394" watchObservedRunningTime="2025-12-08 14:49:20.027153931 +0000 UTC m=+181.127160046" Dec 08 14:49:20 crc kubenswrapper[4894]: I1208 14:49:20.968381 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tglg6" event={"ID":"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1","Type":"ContainerStarted","Data":"ebf1654d54716f7f6a0e9f0397a7f424207fd7b07f06af4525da2492326620ab"} Dec 08 14:49:20 crc kubenswrapper[4894]: I1208 14:49:20.988767 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tglg6" podStartSLOduration=2.160356491 podStartE2EDuration="34.988746126s" podCreationTimestamp="2025-12-08 14:48:46 +0000 UTC" firstStartedPulling="2025-12-08 14:48:47.621579442 +0000 UTC m=+148.721585557" lastFinishedPulling="2025-12-08 14:49:20.449969077 +0000 UTC m=+181.549975192" observedRunningTime="2025-12-08 14:49:20.986308413 +0000 UTC m=+182.086314538" watchObservedRunningTime="2025-12-08 14:49:20.988746126 +0000 UTC m=+182.088752241" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.344508 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 08 14:49:22 crc kubenswrapper[4894]: E1208 14:49:22.345207 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42ba905c-6dd6-41c1-8703-6160ee507486" containerName="pruner" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.345221 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="42ba905c-6dd6-41c1-8703-6160ee507486" containerName="pruner" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.345340 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="42ba905c-6dd6-41c1-8703-6160ee507486" containerName="pruner" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.345859 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.349442 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.350877 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.387930 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.451106 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e2d51c-9c99-4fc0-9978-827549f8b918-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.451586 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5e2d51c-9c99-4fc0-9978-827549f8b918-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.552955 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e2d51c-9c99-4fc0-9978-827549f8b918-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.553095 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5e2d51c-9c99-4fc0-9978-827549f8b918-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.553185 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5e2d51c-9c99-4fc0-9978-827549f8b918-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.575903 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e2d51c-9c99-4fc0-9978-827549f8b918-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:22 crc kubenswrapper[4894]: I1208 14:49:22.700131 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:23 crc kubenswrapper[4894]: I1208 14:49:23.101206 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 08 14:49:23 crc kubenswrapper[4894]: W1208 14:49:23.106907 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode5e2d51c_9c99_4fc0_9978_827549f8b918.slice/crio-b6e3e9aa4876a0788992a2fca8bb16bf8bd88ebce3987483eb709945db494017 WatchSource:0}: Error finding container b6e3e9aa4876a0788992a2fca8bb16bf8bd88ebce3987483eb709945db494017: Status 404 returned error can't find the container with id b6e3e9aa4876a0788992a2fca8bb16bf8bd88ebce3987483eb709945db494017 Dec 08 14:49:23 crc kubenswrapper[4894]: I1208 14:49:23.986414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e5e2d51c-9c99-4fc0-9978-827549f8b918","Type":"ContainerStarted","Data":"e11a797927801db07d89f1ba12dfa31ce576d9e30e9da81fca082b65a576ebb5"} Dec 08 14:49:23 crc kubenswrapper[4894]: I1208 14:49:23.986751 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e5e2d51c-9c99-4fc0-9978-827549f8b918","Type":"ContainerStarted","Data":"b6e3e9aa4876a0788992a2fca8bb16bf8bd88ebce3987483eb709945db494017"} Dec 08 14:49:23 crc kubenswrapper[4894]: I1208 14:49:23.999683 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.999649259 podStartE2EDuration="1.999649259s" podCreationTimestamp="2025-12-08 14:49:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:49:23.999496864 +0000 UTC m=+185.099503009" watchObservedRunningTime="2025-12-08 14:49:23.999649259 +0000 UTC m=+185.099655374" Dec 08 14:49:24 crc kubenswrapper[4894]: I1208 14:49:24.282425 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xlkww"] Dec 08 14:49:24 crc kubenswrapper[4894]: I1208 14:49:24.994311 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5e2d51c-9c99-4fc0-9978-827549f8b918" containerID="e11a797927801db07d89f1ba12dfa31ce576d9e30e9da81fca082b65a576ebb5" exitCode=0 Dec 08 14:49:24 crc kubenswrapper[4894]: I1208 14:49:24.994341 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e5e2d51c-9c99-4fc0-9978-827549f8b918","Type":"ContainerDied","Data":"e11a797927801db07d89f1ba12dfa31ce576d9e30e9da81fca082b65a576ebb5"} Dec 08 14:49:25 crc kubenswrapper[4894]: I1208 14:49:25.332371 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.365598 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.453455 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.453496 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.509877 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e2d51c-9c99-4fc0-9978-827549f8b918-kube-api-access\") pod \"e5e2d51c-9c99-4fc0-9978-827549f8b918\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.509995 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5e2d51c-9c99-4fc0-9978-827549f8b918-kubelet-dir\") pod \"e5e2d51c-9c99-4fc0-9978-827549f8b918\" (UID: \"e5e2d51c-9c99-4fc0-9978-827549f8b918\") " Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.510118 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e5e2d51c-9c99-4fc0-9978-827549f8b918-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e5e2d51c-9c99-4fc0-9978-827549f8b918" (UID: "e5e2d51c-9c99-4fc0-9978-827549f8b918"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.510415 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e5e2d51c-9c99-4fc0-9978-827549f8b918-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.515519 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.517156 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5e2d51c-9c99-4fc0-9978-827549f8b918-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e5e2d51c-9c99-4fc0-9978-827549f8b918" (UID: "e5e2d51c-9c99-4fc0-9978-827549f8b918"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:49:26 crc kubenswrapper[4894]: I1208 14:49:26.611246 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e5e2d51c-9c99-4fc0-9978-827549f8b918-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:27 crc kubenswrapper[4894]: I1208 14:49:27.008609 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e5e2d51c-9c99-4fc0-9978-827549f8b918","Type":"ContainerDied","Data":"b6e3e9aa4876a0788992a2fca8bb16bf8bd88ebce3987483eb709945db494017"} Dec 08 14:49:27 crc kubenswrapper[4894]: I1208 14:49:27.008658 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 08 14:49:27 crc kubenswrapper[4894]: I1208 14:49:27.008671 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6e3e9aa4876a0788992a2fca8bb16bf8bd88ebce3987483eb709945db494017" Dec 08 14:49:27 crc kubenswrapper[4894]: I1208 14:49:27.075998 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:49:27 crc kubenswrapper[4894]: I1208 14:49:27.127382 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tglg6"] Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.020563 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tglg6" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="registry-server" containerID="cri-o://ebf1654d54716f7f6a0e9f0397a7f424207fd7b07f06af4525da2492326620ab" gracePeriod=2 Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.935087 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 08 14:49:29 crc kubenswrapper[4894]: E1208 14:49:29.935651 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e2d51c-9c99-4fc0-9978-827549f8b918" containerName="pruner" Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.935666 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e2d51c-9c99-4fc0-9978-827549f8b918" containerName="pruner" Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.935765 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5e2d51c-9c99-4fc0-9978-827549f8b918" containerName="pruner" Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.936272 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.938593 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.939619 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 08 14:49:29 crc kubenswrapper[4894]: I1208 14:49:29.945678 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.030701 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjtc9" event={"ID":"0edc032d-9fd8-4fbf-b7c7-9106e78404ee","Type":"ContainerStarted","Data":"e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca"} Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.033462 4894 generic.go:334] "Generic (PLEG): container finished" podID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerID="ebf1654d54716f7f6a0e9f0397a7f424207fd7b07f06af4525da2492326620ab" exitCode=0 Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.033518 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tglg6" event={"ID":"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1","Type":"ContainerDied","Data":"ebf1654d54716f7f6a0e9f0397a7f424207fd7b07f06af4525da2492326620ab"} Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.053945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-var-lock\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.054016 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.054092 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40041abc-c8e0-46d9-acdc-c4789f020b6c-kube-api-access\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.154743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-var-lock\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.154801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.154873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40041abc-c8e0-46d9-acdc-c4789f020b6c-kube-api-access\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.154887 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-var-lock\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.154897 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.176195 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40041abc-c8e0-46d9-acdc-c4789f020b6c-kube-api-access\") pod \"installer-9-crc\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.254619 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:49:30 crc kubenswrapper[4894]: I1208 14:49:30.648630 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 08 14:49:30 crc kubenswrapper[4894]: W1208 14:49:30.676794 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod40041abc_c8e0_46d9_acdc_c4789f020b6c.slice/crio-d3873d4167492ee6797004a9554f0f39a08da6ecea7b776ad6ee52f1a8a062bc WatchSource:0}: Error finding container d3873d4167492ee6797004a9554f0f39a08da6ecea7b776ad6ee52f1a8a062bc: Status 404 returned error can't find the container with id d3873d4167492ee6797004a9554f0f39a08da6ecea7b776ad6ee52f1a8a062bc Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.049299 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40041abc-c8e0-46d9-acdc-c4789f020b6c","Type":"ContainerStarted","Data":"d3873d4167492ee6797004a9554f0f39a08da6ecea7b776ad6ee52f1a8a062bc"} Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.051340 4894 generic.go:334] "Generic (PLEG): container finished" podID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerID="e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca" exitCode=0 Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.051378 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjtc9" event={"ID":"0edc032d-9fd8-4fbf-b7c7-9106e78404ee","Type":"ContainerDied","Data":"e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca"} Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.268660 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.467792 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-catalog-content\") pod \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.467949 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nbzv\" (UniqueName: \"kubernetes.io/projected/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-kube-api-access-6nbzv\") pod \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.467978 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-utilities\") pod \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\" (UID: \"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1\") " Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.468720 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-utilities" (OuterVolumeSpecName: "utilities") pod "f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" (UID: "f33c42c6-d4ea-4b9c-82fc-453810ab1bb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.477082 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-kube-api-access-6nbzv" (OuterVolumeSpecName: "kube-api-access-6nbzv") pod "f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" (UID: "f33c42c6-d4ea-4b9c-82fc-453810ab1bb1"). InnerVolumeSpecName "kube-api-access-6nbzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.569187 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nbzv\" (UniqueName: \"kubernetes.io/projected/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-kube-api-access-6nbzv\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.569237 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.582981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" (UID: "f33c42c6-d4ea-4b9c-82fc-453810ab1bb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:31 crc kubenswrapper[4894]: I1208 14:49:31.672493 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.059978 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tglg6" event={"ID":"f33c42c6-d4ea-4b9c-82fc-453810ab1bb1","Type":"ContainerDied","Data":"193db2d821d18104791d5b683870e4efac1027ac25948221a7b3aa6bfe654d4b"} Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.060039 4894 scope.go:117] "RemoveContainer" containerID="ebf1654d54716f7f6a0e9f0397a7f424207fd7b07f06af4525da2492326620ab" Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.060193 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tglg6" Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.072322 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zntsv" event={"ID":"4efaa068-5a66-4fbc-902d-1d535089a121","Type":"ContainerStarted","Data":"1e2689a442052b5699dfa8b2fadc8b05f5d52bc6a45da70b89dcca9cbfb0ef97"} Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.075088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbbpn" event={"ID":"a4d8a8de-7e5f-4387-8a9f-281421db9641","Type":"ContainerStarted","Data":"cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab"} Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.077938 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40041abc-c8e0-46d9-acdc-c4789f020b6c","Type":"ContainerStarted","Data":"26a7749776c21c830deccf4d5a0b2477e150c53d6302a393f123a258a6109ee2"} Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.093535 4894 scope.go:117] "RemoveContainer" containerID="cef07a9e0a10da78165c7115d227f0de203dce4e6433af592118abf805f27d81" Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.103746 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tglg6"] Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.109627 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tglg6"] Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.120140 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.120122512 podStartE2EDuration="3.120122512s" podCreationTimestamp="2025-12-08 14:49:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:49:32.119420349 +0000 UTC m=+193.219426464" watchObservedRunningTime="2025-12-08 14:49:32.120122512 +0000 UTC m=+193.220128627" Dec 08 14:49:32 crc kubenswrapper[4894]: I1208 14:49:32.129934 4894 scope.go:117] "RemoveContainer" containerID="d039f2a346e669e9d9bf10cb85802a4847ac842ea2447ab2f2490c71e33ac212" Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.085477 4894 generic.go:334] "Generic (PLEG): container finished" podID="4efaa068-5a66-4fbc-902d-1d535089a121" containerID="1e2689a442052b5699dfa8b2fadc8b05f5d52bc6a45da70b89dcca9cbfb0ef97" exitCode=0 Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.085572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zntsv" event={"ID":"4efaa068-5a66-4fbc-902d-1d535089a121","Type":"ContainerDied","Data":"1e2689a442052b5699dfa8b2fadc8b05f5d52bc6a45da70b89dcca9cbfb0ef97"} Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.088485 4894 generic.go:334] "Generic (PLEG): container finished" podID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerID="c6d12f825b5eff89b63364b7185427fef732824c1530f0e56673a37f97415d21" exitCode=0 Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.088612 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65wft" event={"ID":"07738ca6-502f-43d9-a3bc-951b99d0a0fc","Type":"ContainerDied","Data":"c6d12f825b5eff89b63364b7185427fef732824c1530f0e56673a37f97415d21"} Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.090418 4894 generic.go:334] "Generic (PLEG): container finished" podID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerID="cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab" exitCode=0 Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.090475 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbbpn" event={"ID":"a4d8a8de-7e5f-4387-8a9f-281421db9641","Type":"ContainerDied","Data":"cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab"} Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.096289 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjtc9" event={"ID":"0edc032d-9fd8-4fbf-b7c7-9106e78404ee","Type":"ContainerStarted","Data":"b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287"} Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.159012 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xjtc9" podStartSLOduration=3.321929445 podStartE2EDuration="51.158989232s" podCreationTimestamp="2025-12-08 14:48:42 +0000 UTC" firstStartedPulling="2025-12-08 14:48:44.478713005 +0000 UTC m=+145.578719120" lastFinishedPulling="2025-12-08 14:49:32.315772802 +0000 UTC m=+193.415778907" observedRunningTime="2025-12-08 14:49:33.158174094 +0000 UTC m=+194.258180209" watchObservedRunningTime="2025-12-08 14:49:33.158989232 +0000 UTC m=+194.258995347" Dec 08 14:49:33 crc kubenswrapper[4894]: I1208 14:49:33.204523 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" path="/var/lib/kubelet/pods/f33c42c6-d4ea-4b9c-82fc-453810ab1bb1/volumes" Dec 08 14:49:34 crc kubenswrapper[4894]: I1208 14:49:34.104410 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zntsv" event={"ID":"4efaa068-5a66-4fbc-902d-1d535089a121","Type":"ContainerStarted","Data":"6025c1a37a4437634b0a03578ebb63b1314a58d3d01f6c307a3389ab9feea4eb"} Dec 08 14:49:34 crc kubenswrapper[4894]: I1208 14:49:34.106354 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65wft" event={"ID":"07738ca6-502f-43d9-a3bc-951b99d0a0fc","Type":"ContainerStarted","Data":"ade5c65d06f1ed792800971dea967cb5fe9f3c5226c7219dd90303262ee651d4"} Dec 08 14:49:34 crc kubenswrapper[4894]: I1208 14:49:34.107999 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbbpn" event={"ID":"a4d8a8de-7e5f-4387-8a9f-281421db9641","Type":"ContainerStarted","Data":"da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8"} Dec 08 14:49:34 crc kubenswrapper[4894]: I1208 14:49:34.157831 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-65wft" podStartSLOduration=4.166355705 podStartE2EDuration="52.157794037s" podCreationTimestamp="2025-12-08 14:48:42 +0000 UTC" firstStartedPulling="2025-12-08 14:48:45.541524465 +0000 UTC m=+146.641530580" lastFinishedPulling="2025-12-08 14:49:33.532962797 +0000 UTC m=+194.632968912" observedRunningTime="2025-12-08 14:49:34.154271328 +0000 UTC m=+195.254277443" watchObservedRunningTime="2025-12-08 14:49:34.157794037 +0000 UTC m=+195.257800172" Dec 08 14:49:34 crc kubenswrapper[4894]: I1208 14:49:34.159340 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zntsv" podStartSLOduration=4.188337759 podStartE2EDuration="52.159330729s" podCreationTimestamp="2025-12-08 14:48:42 +0000 UTC" firstStartedPulling="2025-12-08 14:48:45.503171547 +0000 UTC m=+146.603177662" lastFinishedPulling="2025-12-08 14:49:33.474164517 +0000 UTC m=+194.574170632" observedRunningTime="2025-12-08 14:49:34.132405268 +0000 UTC m=+195.232411383" watchObservedRunningTime="2025-12-08 14:49:34.159330729 +0000 UTC m=+195.259336844" Dec 08 14:49:34 crc kubenswrapper[4894]: I1208 14:49:34.175196 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gbbpn" podStartSLOduration=2.233564834 podStartE2EDuration="49.175173715s" podCreationTimestamp="2025-12-08 14:48:45 +0000 UTC" firstStartedPulling="2025-12-08 14:48:46.56284059 +0000 UTC m=+147.662846705" lastFinishedPulling="2025-12-08 14:49:33.504449471 +0000 UTC m=+194.604455586" observedRunningTime="2025-12-08 14:49:34.168672836 +0000 UTC m=+195.268678951" watchObservedRunningTime="2025-12-08 14:49:34.175173715 +0000 UTC m=+195.275179830" Dec 08 14:49:35 crc kubenswrapper[4894]: I1208 14:49:35.115961 4894 generic.go:334] "Generic (PLEG): container finished" podID="ccba7943-00e3-422a-91a6-a85439a4c174" containerID="dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343" exitCode=0 Dec 08 14:49:35 crc kubenswrapper[4894]: I1208 14:49:35.116141 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4hbv" event={"ID":"ccba7943-00e3-422a-91a6-a85439a4c174","Type":"ContainerDied","Data":"dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343"} Dec 08 14:49:35 crc kubenswrapper[4894]: I1208 14:49:35.969749 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:49:35 crc kubenswrapper[4894]: I1208 14:49:35.970147 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:49:36 crc kubenswrapper[4894]: I1208 14:49:36.123177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4hbv" event={"ID":"ccba7943-00e3-422a-91a6-a85439a4c174","Type":"ContainerStarted","Data":"6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903"} Dec 08 14:49:36 crc kubenswrapper[4894]: I1208 14:49:36.143079 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h4hbv" podStartSLOduration=1.827721517 podStartE2EDuration="52.143060548s" podCreationTimestamp="2025-12-08 14:48:44 +0000 UTC" firstStartedPulling="2025-12-08 14:48:45.49057059 +0000 UTC m=+146.590576705" lastFinishedPulling="2025-12-08 14:49:35.805909621 +0000 UTC m=+196.905915736" observedRunningTime="2025-12-08 14:49:36.13955828 +0000 UTC m=+197.239564395" watchObservedRunningTime="2025-12-08 14:49:36.143060548 +0000 UTC m=+197.243066663" Dec 08 14:49:37 crc kubenswrapper[4894]: I1208 14:49:37.013152 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gbbpn" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="registry-server" probeResult="failure" output=< Dec 08 14:49:37 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 08 14:49:37 crc kubenswrapper[4894]: > Dec 08 14:49:37 crc kubenswrapper[4894]: I1208 14:49:37.130601 4894 generic.go:334] "Generic (PLEG): container finished" podID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerID="8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448" exitCode=0 Dec 08 14:49:37 crc kubenswrapper[4894]: I1208 14:49:37.130638 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzlrr" event={"ID":"17e3e7c4-52dd-48d3-b819-a0fed778803b","Type":"ContainerDied","Data":"8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448"} Dec 08 14:49:37 crc kubenswrapper[4894]: I1208 14:49:37.133438 4894 generic.go:334] "Generic (PLEG): container finished" podID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerID="95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914" exitCode=0 Dec 08 14:49:37 crc kubenswrapper[4894]: I1208 14:49:37.133526 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t69pb" event={"ID":"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e","Type":"ContainerDied","Data":"95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914"} Dec 08 14:49:37 crc kubenswrapper[4894]: I1208 14:49:37.296800 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:49:37 crc kubenswrapper[4894]: I1208 14:49:37.296900 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:49:40 crc kubenswrapper[4894]: I1208 14:49:40.154314 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzlrr" event={"ID":"17e3e7c4-52dd-48d3-b819-a0fed778803b","Type":"ContainerStarted","Data":"2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb"} Dec 08 14:49:40 crc kubenswrapper[4894]: I1208 14:49:40.157171 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t69pb" event={"ID":"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e","Type":"ContainerStarted","Data":"a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea"} Dec 08 14:49:40 crc kubenswrapper[4894]: I1208 14:49:40.172001 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jzlrr" podStartSLOduration=2.979626354 podStartE2EDuration="56.171969327s" podCreationTimestamp="2025-12-08 14:48:44 +0000 UTC" firstStartedPulling="2025-12-08 14:48:46.557537751 +0000 UTC m=+147.657543866" lastFinishedPulling="2025-12-08 14:49:39.749880724 +0000 UTC m=+200.849886839" observedRunningTime="2025-12-08 14:49:40.171031476 +0000 UTC m=+201.271037601" watchObservedRunningTime="2025-12-08 14:49:40.171969327 +0000 UTC m=+201.271975442" Dec 08 14:49:40 crc kubenswrapper[4894]: I1208 14:49:40.189044 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t69pb" podStartSLOduration=2.7917305150000002 podStartE2EDuration="58.189026097s" podCreationTimestamp="2025-12-08 14:48:42 +0000 UTC" firstStartedPulling="2025-12-08 14:48:44.362612256 +0000 UTC m=+145.462618381" lastFinishedPulling="2025-12-08 14:49:39.759907848 +0000 UTC m=+200.859913963" observedRunningTime="2025-12-08 14:49:40.187872518 +0000 UTC m=+201.287878643" watchObservedRunningTime="2025-12-08 14:49:40.189026097 +0000 UTC m=+201.289032212" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.391880 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.392159 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.434699 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.641567 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.641660 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.685015 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.793665 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.793735 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.832028 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.989263 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:49:42 crc kubenswrapper[4894]: I1208 14:49:42.989616 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:49:43 crc kubenswrapper[4894]: I1208 14:49:43.035725 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:49:43 crc kubenswrapper[4894]: I1208 14:49:43.229478 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:49:43 crc kubenswrapper[4894]: I1208 14:49:43.229879 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:49:43 crc kubenswrapper[4894]: I1208 14:49:43.255618 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:49:44 crc kubenswrapper[4894]: I1208 14:49:44.251630 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:49:44 crc kubenswrapper[4894]: I1208 14:49:44.397318 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:49:44 crc kubenswrapper[4894]: I1208 14:49:44.397388 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:49:44 crc kubenswrapper[4894]: I1208 14:49:44.440167 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:49:44 crc kubenswrapper[4894]: I1208 14:49:44.781635 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:49:44 crc kubenswrapper[4894]: I1208 14:49:44.781975 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:49:44 crc kubenswrapper[4894]: I1208 14:49:44.816290 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:49:45 crc kubenswrapper[4894]: I1208 14:49:45.235717 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:49:45 crc kubenswrapper[4894]: I1208 14:49:45.267737 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:49:45 crc kubenswrapper[4894]: I1208 14:49:45.342296 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zntsv"] Dec 08 14:49:45 crc kubenswrapper[4894]: I1208 14:49:45.342558 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zntsv" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="registry-server" containerID="cri-o://6025c1a37a4437634b0a03578ebb63b1314a58d3d01f6c307a3389ab9feea4eb" gracePeriod=2 Dec 08 14:49:45 crc kubenswrapper[4894]: I1208 14:49:45.544156 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-65wft"] Dec 08 14:49:45 crc kubenswrapper[4894]: I1208 14:49:45.544468 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-65wft" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="registry-server" containerID="cri-o://ade5c65d06f1ed792800971dea967cb5fe9f3c5226c7219dd90303262ee651d4" gracePeriod=2 Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.009486 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.050738 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.207549 4894 generic.go:334] "Generic (PLEG): container finished" podID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerID="ade5c65d06f1ed792800971dea967cb5fe9f3c5226c7219dd90303262ee651d4" exitCode=0 Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.207612 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65wft" event={"ID":"07738ca6-502f-43d9-a3bc-951b99d0a0fc","Type":"ContainerDied","Data":"ade5c65d06f1ed792800971dea967cb5fe9f3c5226c7219dd90303262ee651d4"} Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.210287 4894 generic.go:334] "Generic (PLEG): container finished" podID="4efaa068-5a66-4fbc-902d-1d535089a121" containerID="6025c1a37a4437634b0a03578ebb63b1314a58d3d01f6c307a3389ab9feea4eb" exitCode=0 Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.210336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zntsv" event={"ID":"4efaa068-5a66-4fbc-902d-1d535089a121","Type":"ContainerDied","Data":"6025c1a37a4437634b0a03578ebb63b1314a58d3d01f6c307a3389ab9feea4eb"} Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.262048 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.362288 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66cmn\" (UniqueName: \"kubernetes.io/projected/4efaa068-5a66-4fbc-902d-1d535089a121-kube-api-access-66cmn\") pod \"4efaa068-5a66-4fbc-902d-1d535089a121\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.362445 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-catalog-content\") pod \"4efaa068-5a66-4fbc-902d-1d535089a121\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.362481 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-utilities\") pod \"4efaa068-5a66-4fbc-902d-1d535089a121\" (UID: \"4efaa068-5a66-4fbc-902d-1d535089a121\") " Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.363402 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-utilities" (OuterVolumeSpecName: "utilities") pod "4efaa068-5a66-4fbc-902d-1d535089a121" (UID: "4efaa068-5a66-4fbc-902d-1d535089a121"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.364029 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.367722 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4efaa068-5a66-4fbc-902d-1d535089a121-kube-api-access-66cmn" (OuterVolumeSpecName: "kube-api-access-66cmn") pod "4efaa068-5a66-4fbc-902d-1d535089a121" (UID: "4efaa068-5a66-4fbc-902d-1d535089a121"). InnerVolumeSpecName "kube-api-access-66cmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.418725 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4efaa068-5a66-4fbc-902d-1d535089a121" (UID: "4efaa068-5a66-4fbc-902d-1d535089a121"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.465646 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4efaa068-5a66-4fbc-902d-1d535089a121-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.465682 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66cmn\" (UniqueName: \"kubernetes.io/projected/4efaa068-5a66-4fbc-902d-1d535089a121-kube-api-access-66cmn\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.572418 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.768462 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-catalog-content\") pod \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.768519 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmtnw\" (UniqueName: \"kubernetes.io/projected/07738ca6-502f-43d9-a3bc-951b99d0a0fc-kube-api-access-qmtnw\") pod \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.768574 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-utilities\") pod \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\" (UID: \"07738ca6-502f-43d9-a3bc-951b99d0a0fc\") " Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.769436 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-utilities" (OuterVolumeSpecName: "utilities") pod "07738ca6-502f-43d9-a3bc-951b99d0a0fc" (UID: "07738ca6-502f-43d9-a3bc-951b99d0a0fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.771056 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07738ca6-502f-43d9-a3bc-951b99d0a0fc-kube-api-access-qmtnw" (OuterVolumeSpecName: "kube-api-access-qmtnw") pod "07738ca6-502f-43d9-a3bc-951b99d0a0fc" (UID: "07738ca6-502f-43d9-a3bc-951b99d0a0fc"). InnerVolumeSpecName "kube-api-access-qmtnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.817525 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07738ca6-502f-43d9-a3bc-951b99d0a0fc" (UID: "07738ca6-502f-43d9-a3bc-951b99d0a0fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.870711 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.871109 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmtnw\" (UniqueName: \"kubernetes.io/projected/07738ca6-502f-43d9-a3bc-951b99d0a0fc-kube-api-access-qmtnw\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:46 crc kubenswrapper[4894]: I1208 14:49:46.871211 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07738ca6-502f-43d9-a3bc-951b99d0a0fc-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.219534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zntsv" event={"ID":"4efaa068-5a66-4fbc-902d-1d535089a121","Type":"ContainerDied","Data":"d0731ab87636cdde1fd0958320b641b093fde856a951256bd10bc3eebde420f9"} Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.219632 4894 scope.go:117] "RemoveContainer" containerID="6025c1a37a4437634b0a03578ebb63b1314a58d3d01f6c307a3389ab9feea4eb" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.219996 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zntsv" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.224056 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65wft" event={"ID":"07738ca6-502f-43d9-a3bc-951b99d0a0fc","Type":"ContainerDied","Data":"61edb2f22078539a18c1d09a30263584daca09a671e2e48f15f317c1bd8a3bae"} Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.224209 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65wft" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.245803 4894 scope.go:117] "RemoveContainer" containerID="1e2689a442052b5699dfa8b2fadc8b05f5d52bc6a45da70b89dcca9cbfb0ef97" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.252510 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-65wft"] Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.255359 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-65wft"] Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.272106 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zntsv"] Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.275221 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zntsv"] Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.280632 4894 scope.go:117] "RemoveContainer" containerID="7a8bc8bb8edaaeaf0fbf6eeaecdeabd21258334f4416716c4718c7df17a30934" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.296115 4894 scope.go:117] "RemoveContainer" containerID="ade5c65d06f1ed792800971dea967cb5fe9f3c5226c7219dd90303262ee651d4" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.314985 4894 scope.go:117] "RemoveContainer" containerID="c6d12f825b5eff89b63364b7185427fef732824c1530f0e56673a37f97415d21" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.338512 4894 scope.go:117] "RemoveContainer" containerID="b84136b3cd76252e4ce31d0e91a826ba2d9de867aa226518723fd0957d49452e" Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.743506 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzlrr"] Dec 08 14:49:47 crc kubenswrapper[4894]: I1208 14:49:47.744096 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jzlrr" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="registry-server" containerID="cri-o://2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb" gracePeriod=2 Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.600196 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.697200 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-catalog-content\") pod \"17e3e7c4-52dd-48d3-b819-a0fed778803b\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.697250 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-utilities\") pod \"17e3e7c4-52dd-48d3-b819-a0fed778803b\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.697273 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kt7h\" (UniqueName: \"kubernetes.io/projected/17e3e7c4-52dd-48d3-b819-a0fed778803b-kube-api-access-6kt7h\") pod \"17e3e7c4-52dd-48d3-b819-a0fed778803b\" (UID: \"17e3e7c4-52dd-48d3-b819-a0fed778803b\") " Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.698143 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-utilities" (OuterVolumeSpecName: "utilities") pod "17e3e7c4-52dd-48d3-b819-a0fed778803b" (UID: "17e3e7c4-52dd-48d3-b819-a0fed778803b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.702330 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e3e7c4-52dd-48d3-b819-a0fed778803b-kube-api-access-6kt7h" (OuterVolumeSpecName: "kube-api-access-6kt7h") pod "17e3e7c4-52dd-48d3-b819-a0fed778803b" (UID: "17e3e7c4-52dd-48d3-b819-a0fed778803b"). InnerVolumeSpecName "kube-api-access-6kt7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.718533 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17e3e7c4-52dd-48d3-b819-a0fed778803b" (UID: "17e3e7c4-52dd-48d3-b819-a0fed778803b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.798048 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.798080 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17e3e7c4-52dd-48d3-b819-a0fed778803b-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:48 crc kubenswrapper[4894]: I1208 14:49:48.798090 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kt7h\" (UniqueName: \"kubernetes.io/projected/17e3e7c4-52dd-48d3-b819-a0fed778803b-kube-api-access-6kt7h\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.204025 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" path="/var/lib/kubelet/pods/07738ca6-502f-43d9-a3bc-951b99d0a0fc/volumes" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.204717 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" path="/var/lib/kubelet/pods/4efaa068-5a66-4fbc-902d-1d535089a121/volumes" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.245200 4894 generic.go:334] "Generic (PLEG): container finished" podID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerID="2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb" exitCode=0 Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.245249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzlrr" event={"ID":"17e3e7c4-52dd-48d3-b819-a0fed778803b","Type":"ContainerDied","Data":"2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb"} Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.245280 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jzlrr" event={"ID":"17e3e7c4-52dd-48d3-b819-a0fed778803b","Type":"ContainerDied","Data":"7b52c1b2ba9e2c592106de50f57531fe5e4faeddb589cf52d528dd37badd604b"} Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.245301 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jzlrr" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.245316 4894 scope.go:117] "RemoveContainer" containerID="2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.268372 4894 scope.go:117] "RemoveContainer" containerID="8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.281397 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzlrr"] Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.288090 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jzlrr"] Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.297642 4894 scope.go:117] "RemoveContainer" containerID="550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.307995 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" podUID="8cdd7414-21ba-4649-9514-66f2e3688022" containerName="oauth-openshift" containerID="cri-o://fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf" gracePeriod=15 Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.313495 4894 scope.go:117] "RemoveContainer" containerID="2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb" Dec 08 14:49:49 crc kubenswrapper[4894]: E1208 14:49:49.314359 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb\": container with ID starting with 2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb not found: ID does not exist" containerID="2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.314421 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb"} err="failed to get container status \"2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb\": rpc error: code = NotFound desc = could not find container \"2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb\": container with ID starting with 2b10c22b05247d23766fc69cd8db4ed9fcb5c9f5c7d59c764ec49f402101d6bb not found: ID does not exist" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.314504 4894 scope.go:117] "RemoveContainer" containerID="8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448" Dec 08 14:49:49 crc kubenswrapper[4894]: E1208 14:49:49.322563 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448\": container with ID starting with 8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448 not found: ID does not exist" containerID="8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.322634 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448"} err="failed to get container status \"8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448\": rpc error: code = NotFound desc = could not find container \"8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448\": container with ID starting with 8d74885db0b31ff3653a6dab5165a320cd148a4f091aac9fc806859d76527448 not found: ID does not exist" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.322680 4894 scope.go:117] "RemoveContainer" containerID="550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413" Dec 08 14:49:49 crc kubenswrapper[4894]: E1208 14:49:49.323385 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413\": container with ID starting with 550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413 not found: ID does not exist" containerID="550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413" Dec 08 14:49:49 crc kubenswrapper[4894]: I1208 14:49:49.323443 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413"} err="failed to get container status \"550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413\": rpc error: code = NotFound desc = could not find container \"550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413\": container with ID starting with 550aebfe67a547bd4e9ddef0342b108f8eca9546cff58e3674de86f90b3d1413 not found: ID does not exist" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.161441 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214691 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-session\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214742 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-trusted-ca-bundle\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214765 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-router-certs\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214784 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-provider-selection\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214835 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-error\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214853 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-service-ca\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214869 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-cliconfig\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214890 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8cdd7414-21ba-4649-9514-66f2e3688022-audit-dir\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214924 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-serving-cert\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214941 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-audit-policies\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214963 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h28nv\" (UniqueName: \"kubernetes.io/projected/8cdd7414-21ba-4649-9514-66f2e3688022-kube-api-access-h28nv\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.214986 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-ocp-branding-template\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.215009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-idp-0-file-data\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.215032 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-login\") pod \"8cdd7414-21ba-4649-9514-66f2e3688022\" (UID: \"8cdd7414-21ba-4649-9514-66f2e3688022\") " Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.215725 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cdd7414-21ba-4649-9514-66f2e3688022-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.216267 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.216374 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.216717 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.218418 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.220196 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.220327 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cdd7414-21ba-4649-9514-66f2e3688022-kube-api-access-h28nv" (OuterVolumeSpecName: "kube-api-access-h28nv") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "kube-api-access-h28nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.221755 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.221939 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.222112 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.222739 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.233105 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.233246 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.233466 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "8cdd7414-21ba-4649-9514-66f2e3688022" (UID: "8cdd7414-21ba-4649-9514-66f2e3688022"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.256544 4894 generic.go:334] "Generic (PLEG): container finished" podID="8cdd7414-21ba-4649-9514-66f2e3688022" containerID="fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf" exitCode=0 Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.256593 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" event={"ID":"8cdd7414-21ba-4649-9514-66f2e3688022","Type":"ContainerDied","Data":"fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf"} Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.256858 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" event={"ID":"8cdd7414-21ba-4649-9514-66f2e3688022","Type":"ContainerDied","Data":"5d394988f9dd91321666eb557307d51d6dead9310806c66824876e64839bc11f"} Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.256891 4894 scope.go:117] "RemoveContainer" containerID="fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.256903 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-xlkww" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.275053 4894 scope.go:117] "RemoveContainer" containerID="fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.282532 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xlkww"] Dec 08 14:49:50 crc kubenswrapper[4894]: E1208 14:49:50.282662 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf\": container with ID starting with fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf not found: ID does not exist" containerID="fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.282706 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf"} err="failed to get container status \"fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf\": rpc error: code = NotFound desc = could not find container \"fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf\": container with ID starting with fece4e3d079fdf9e7df19e147fee0e4d5a15a03fd03399618fed02208a882bdf not found: ID does not exist" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.284712 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-xlkww"] Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.315901 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.315988 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316005 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316017 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316031 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316052 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316081 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316096 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316109 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316122 4894 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8cdd7414-21ba-4649-9514-66f2e3688022-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316133 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316146 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdd7414-21ba-4649-9514-66f2e3688022-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316156 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8cdd7414-21ba-4649-9514-66f2e3688022-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:50 crc kubenswrapper[4894]: I1208 14:49:50.316168 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h28nv\" (UniqueName: \"kubernetes.io/projected/8cdd7414-21ba-4649-9514-66f2e3688022-kube-api-access-h28nv\") on node \"crc\" DevicePath \"\"" Dec 08 14:49:51 crc kubenswrapper[4894]: I1208 14:49:51.207995 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" path="/var/lib/kubelet/pods/17e3e7c4-52dd-48d3-b819-a0fed778803b/volumes" Dec 08 14:49:51 crc kubenswrapper[4894]: I1208 14:49:51.209054 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cdd7414-21ba-4649-9514-66f2e3688022" path="/var/lib/kubelet/pods/8cdd7414-21ba-4649-9514-66f2e3688022/volumes" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.428325 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-59978ccd59-j6qb4"] Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429414 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429435 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429449 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429457 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429471 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429477 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429486 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429493 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429503 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429509 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429519 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429525 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429534 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429540 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429549 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429555 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429564 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429570 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429579 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429584 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="extract-utilities" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429593 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cdd7414-21ba-4649-9514-66f2e3688022" containerName="oauth-openshift" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429599 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cdd7414-21ba-4649-9514-66f2e3688022" containerName="oauth-openshift" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429605 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429610 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: E1208 14:49:54.429623 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429629 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="extract-content" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429738 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4efaa068-5a66-4fbc-902d-1d535089a121" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429748 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cdd7414-21ba-4649-9514-66f2e3688022" containerName="oauth-openshift" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429757 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="07738ca6-502f-43d9-a3bc-951b99d0a0fc" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429766 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e3e7c4-52dd-48d3-b819-a0fed778803b" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.429775 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33c42c6-d4ea-4b9c-82fc-453810ab1bb1" containerName="registry-server" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.430380 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.433988 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.434467 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.435931 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.436360 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.436551 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.436603 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.436701 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.436780 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.437523 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.437693 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.439894 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.440375 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.448406 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.449431 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-59978ccd59-j6qb4"] Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.453151 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.453195 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468415 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-audit-dir\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468468 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468491 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-router-certs\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468508 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-login\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468531 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-session\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468547 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468563 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468580 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468606 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-service-ca\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468866 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xv8j\" (UniqueName: \"kubernetes.io/projected/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-kube-api-access-7xv8j\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.468976 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-error\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.469018 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.469187 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.469228 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-audit-policies\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570106 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xv8j\" (UniqueName: \"kubernetes.io/projected/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-kube-api-access-7xv8j\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-error\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570220 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570273 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-audit-policies\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570370 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-audit-dir\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570429 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-router-certs\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570462 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-login\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570490 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-session\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570516 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570541 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570569 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.570615 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-service-ca\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.573204 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-service-ca\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.575037 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-audit-dir\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.575962 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-audit-policies\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.575997 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.576070 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.579087 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.579196 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.579288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-router-certs\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.579341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-login\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.582265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-session\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.582395 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-user-template-error\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.583969 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.595437 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.600184 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xv8j\" (UniqueName: \"kubernetes.io/projected/c516d94f-00b1-4f92-b49f-f81ca8e74c1c-kube-api-access-7xv8j\") pod \"oauth-openshift-59978ccd59-j6qb4\" (UID: \"c516d94f-00b1-4f92-b49f-f81ca8e74c1c\") " pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:54 crc kubenswrapper[4894]: I1208 14:49:54.747709 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:55 crc kubenswrapper[4894]: I1208 14:49:55.191887 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-59978ccd59-j6qb4"] Dec 08 14:49:55 crc kubenswrapper[4894]: I1208 14:49:55.290467 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" event={"ID":"c516d94f-00b1-4f92-b49f-f81ca8e74c1c","Type":"ContainerStarted","Data":"556cd91487eb1c298798bb129279e389f8750da086270facbc9cb78ea4274e07"} Dec 08 14:49:56 crc kubenswrapper[4894]: I1208 14:49:56.300383 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" event={"ID":"c516d94f-00b1-4f92-b49f-f81ca8e74c1c","Type":"ContainerStarted","Data":"15f330f0c4429e50a60e8c5291a7b821fbfe18b44db972c1f0fb264badc5c13e"} Dec 08 14:49:56 crc kubenswrapper[4894]: I1208 14:49:56.301194 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:56 crc kubenswrapper[4894]: I1208 14:49:56.306121 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" Dec 08 14:49:56 crc kubenswrapper[4894]: I1208 14:49:56.327135 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-59978ccd59-j6qb4" podStartSLOduration=32.327114658 podStartE2EDuration="32.327114658s" podCreationTimestamp="2025-12-08 14:49:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:49:56.325448231 +0000 UTC m=+217.425454356" watchObservedRunningTime="2025-12-08 14:49:56.327114658 +0000 UTC m=+217.427120783" Dec 08 14:50:07 crc kubenswrapper[4894]: I1208 14:50:07.297424 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:50:07 crc kubenswrapper[4894]: I1208 14:50:07.297803 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:50:07 crc kubenswrapper[4894]: I1208 14:50:07.297887 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:50:07 crc kubenswrapper[4894]: I1208 14:50:07.298500 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 14:50:07 crc kubenswrapper[4894]: I1208 14:50:07.298748 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6" gracePeriod=600 Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.363387 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6" exitCode=0 Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.363466 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6"} Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.363716 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"446f34182251fad2ef136856c085177483a1e0666f7e36e3916d271f2f7efad9"} Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.652157 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.652738 4894 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.652879 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.653163 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f" gracePeriod=15 Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.653234 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf" gracePeriod=15 Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.653248 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707" gracePeriod=15 Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.653183 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a" gracePeriod=15 Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.653253 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb" gracePeriod=15 Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.658398 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.658738 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.658754 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.658766 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.658775 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.658785 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.658796 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.658805 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.658876 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.658890 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.658897 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.658913 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.658919 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.659042 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.659056 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.659067 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.659080 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.659089 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.659099 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.659203 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.659211 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 08 14:50:08 crc kubenswrapper[4894]: E1208 14:50:08.689847 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.51:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768356 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768415 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768544 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768583 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768603 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.768639 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.869993 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870056 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870074 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870096 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870165 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870202 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870221 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870419 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870468 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870503 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870527 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870574 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.870622 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:08 crc kubenswrapper[4894]: I1208 14:50:08.990780 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:09 crc kubenswrapper[4894]: E1208 14:50:09.029718 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.51:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f44f790c24895 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-08 14:50:09.028786325 +0000 UTC m=+230.128792440,LastTimestamp:2025-12-08 14:50:09.028786325 +0000 UTC m=+230.128792440,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.200386 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.371593 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2c3103d6ef41104a1e41bd81f9d1ece5755b6ffd376f2770a8b0434c372aa6c3"} Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.373597 4894 generic.go:334] "Generic (PLEG): container finished" podID="40041abc-c8e0-46d9-acdc-c4789f020b6c" containerID="26a7749776c21c830deccf4d5a0b2477e150c53d6302a393f123a258a6109ee2" exitCode=0 Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.373647 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40041abc-c8e0-46d9-acdc-c4789f020b6c","Type":"ContainerDied","Data":"26a7749776c21c830deccf4d5a0b2477e150c53d6302a393f123a258a6109ee2"} Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.374323 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.376625 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.378350 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.379777 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf" exitCode=0 Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.379798 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a" exitCode=0 Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.379807 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb" exitCode=0 Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.379852 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707" exitCode=2 Dec 08 14:50:09 crc kubenswrapper[4894]: I1208 14:50:09.379918 4894 scope.go:117] "RemoveContainer" containerID="8cf934373826fdf77450aa92e09e187ed70d95c6fbee36259ca8b45f6aec90c8" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.386380 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"12a7eda32e24169a241711690e677a618a6f25a64891c7128bfc953dcce265e9"} Dec 08 14:50:10 crc kubenswrapper[4894]: E1208 14:50:10.387044 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.51:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.387117 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.389168 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.605527 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.606179 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.696605 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-kubelet-dir\") pod \"40041abc-c8e0-46d9-acdc-c4789f020b6c\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.696736 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-var-lock\") pod \"40041abc-c8e0-46d9-acdc-c4789f020b6c\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.696828 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40041abc-c8e0-46d9-acdc-c4789f020b6c-kube-api-access\") pod \"40041abc-c8e0-46d9-acdc-c4789f020b6c\" (UID: \"40041abc-c8e0-46d9-acdc-c4789f020b6c\") " Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.696981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-var-lock" (OuterVolumeSpecName: "var-lock") pod "40041abc-c8e0-46d9-acdc-c4789f020b6c" (UID: "40041abc-c8e0-46d9-acdc-c4789f020b6c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.696981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "40041abc-c8e0-46d9-acdc-c4789f020b6c" (UID: "40041abc-c8e0-46d9-acdc-c4789f020b6c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.697081 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.697093 4894 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/40041abc-c8e0-46d9-acdc-c4789f020b6c-var-lock\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.726202 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40041abc-c8e0-46d9-acdc-c4789f020b6c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "40041abc-c8e0-46d9-acdc-c4789f020b6c" (UID: "40041abc-c8e0-46d9-acdc-c4789f020b6c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:50:10 crc kubenswrapper[4894]: I1208 14:50:10.798158 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/40041abc-c8e0-46d9-acdc-c4789f020b6c-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.087826 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.088613 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.089459 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.089644 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.203768 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.203881 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.203941 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.203875 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.203903 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.204058 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.204157 4894 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.204167 4894 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.204178 4894 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.396802 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.397366 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f" exitCode=0 Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.397441 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.397439 4894 scope.go:117] "RemoveContainer" containerID="d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.398174 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.398545 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.399330 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.399395 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"40041abc-c8e0-46d9-acdc-c4789f020b6c","Type":"ContainerDied","Data":"d3873d4167492ee6797004a9554f0f39a08da6ecea7b776ad6ee52f1a8a062bc"} Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.399426 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3873d4167492ee6797004a9554f0f39a08da6ecea7b776ad6ee52f1a8a062bc" Dec 08 14:50:11 crc kubenswrapper[4894]: E1208 14:50:11.399472 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.51:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.402742 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.403170 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.411269 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.411603 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.415551 4894 scope.go:117] "RemoveContainer" containerID="a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.430265 4894 scope.go:117] "RemoveContainer" containerID="a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.444577 4894 scope.go:117] "RemoveContainer" containerID="7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.461720 4894 scope.go:117] "RemoveContainer" containerID="639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.476657 4894 scope.go:117] "RemoveContainer" containerID="4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.495518 4894 scope.go:117] "RemoveContainer" containerID="d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf" Dec 08 14:50:11 crc kubenswrapper[4894]: E1208 14:50:11.496093 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\": container with ID starting with d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf not found: ID does not exist" containerID="d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.496122 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf"} err="failed to get container status \"d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\": rpc error: code = NotFound desc = could not find container \"d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf\": container with ID starting with d88ae1d5df513cfcf45a451e819245830f951a16a64327990ad36c1292af9fdf not found: ID does not exist" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.496143 4894 scope.go:117] "RemoveContainer" containerID="a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a" Dec 08 14:50:11 crc kubenswrapper[4894]: E1208 14:50:11.496388 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\": container with ID starting with a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a not found: ID does not exist" containerID="a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.496413 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a"} err="failed to get container status \"a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\": rpc error: code = NotFound desc = could not find container \"a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a\": container with ID starting with a6c16d5ed41a56c7a1d5dad23bcf0dc7ff01354752c072ec6960c9502af9bb3a not found: ID does not exist" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.496429 4894 scope.go:117] "RemoveContainer" containerID="a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb" Dec 08 14:50:11 crc kubenswrapper[4894]: E1208 14:50:11.496648 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\": container with ID starting with a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb not found: ID does not exist" containerID="a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.496672 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb"} err="failed to get container status \"a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\": rpc error: code = NotFound desc = could not find container \"a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb\": container with ID starting with a35cba52474f0fdad54a304b5eab87b7d81797e4d8dd93e25e9e5726a3d479fb not found: ID does not exist" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.496720 4894 scope.go:117] "RemoveContainer" containerID="7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707" Dec 08 14:50:11 crc kubenswrapper[4894]: E1208 14:50:11.498754 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\": container with ID starting with 7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707 not found: ID does not exist" containerID="7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.498782 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707"} err="failed to get container status \"7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\": rpc error: code = NotFound desc = could not find container \"7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707\": container with ID starting with 7474c58f7819f552b15d8d4efbcbc502bd9095147f47c1694ba2d46c400bc707 not found: ID does not exist" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.498799 4894 scope.go:117] "RemoveContainer" containerID="639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f" Dec 08 14:50:11 crc kubenswrapper[4894]: E1208 14:50:11.499063 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\": container with ID starting with 639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f not found: ID does not exist" containerID="639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.499088 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f"} err="failed to get container status \"639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\": rpc error: code = NotFound desc = could not find container \"639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f\": container with ID starting with 639644ef656598ee1312983c644882dffc0d694f9ca8be90b75f98b8d0b3eb8f not found: ID does not exist" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.499104 4894 scope.go:117] "RemoveContainer" containerID="4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971" Dec 08 14:50:11 crc kubenswrapper[4894]: E1208 14:50:11.499499 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\": container with ID starting with 4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971 not found: ID does not exist" containerID="4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971" Dec 08 14:50:11 crc kubenswrapper[4894]: I1208 14:50:11.499521 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971"} err="failed to get container status \"4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\": rpc error: code = NotFound desc = could not find container \"4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971\": container with ID starting with 4ccc4628d5eaa30d37358562bec79a598b91c195478078b1c80f44bf24128971 not found: ID does not exist" Dec 08 14:50:13 crc kubenswrapper[4894]: I1208 14:50:13.202543 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.054378 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.054871 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.055532 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.055915 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.056142 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:14 crc kubenswrapper[4894]: I1208 14:50:14.056169 4894 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.056344 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="200ms" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.257172 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="400ms" Dec 08 14:50:14 crc kubenswrapper[4894]: E1208 14:50:14.657834 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="800ms" Dec 08 14:50:15 crc kubenswrapper[4894]: E1208 14:50:15.465447 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="1.6s" Dec 08 14:50:17 crc kubenswrapper[4894]: E1208 14:50:17.066250 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="3.2s" Dec 08 14:50:17 crc kubenswrapper[4894]: E1208 14:50:17.492573 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.51:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f44f790c24895 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-08 14:50:09.028786325 +0000 UTC m=+230.128792440,LastTimestamp:2025-12-08 14:50:09.028786325 +0000 UTC m=+230.128792440,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 08 14:50:19 crc kubenswrapper[4894]: I1208 14:50:19.201454 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:19 crc kubenswrapper[4894]: E1208 14:50:19.282512 4894 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.51:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" volumeName="registry-storage" Dec 08 14:50:20 crc kubenswrapper[4894]: E1208 14:50:20.267503 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.51:6443: connect: connection refused" interval="6.4s" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.196485 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.197397 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.209534 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.209570 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:21 crc kubenswrapper[4894]: E1208 14:50:21.210025 4894 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.210521 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:21 crc kubenswrapper[4894]: W1208 14:50:21.229642 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-d1e817d346d72ed86377ada82cf64dfa3e8e18e7e0822fa44a4c403155d8dcda WatchSource:0}: Error finding container d1e817d346d72ed86377ada82cf64dfa3e8e18e7e0822fa44a4c403155d8dcda: Status 404 returned error can't find the container with id d1e817d346d72ed86377ada82cf64dfa3e8e18e7e0822fa44a4c403155d8dcda Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.452404 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1b8171043a837f038d258ca7333e2ef0701937ac7b27863cf85f0a3832921686"} Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.452725 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d1e817d346d72ed86377ada82cf64dfa3e8e18e7e0822fa44a4c403155d8dcda"} Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.455980 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.456041 4894 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169" exitCode=1 Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.456075 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169"} Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.456547 4894 scope.go:117] "RemoveContainer" containerID="342d1c94d596a0ea970e75180c048f3e1b28b2aca93f029ae55dc721ebdad169" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.456916 4894 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.457376 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:21 crc kubenswrapper[4894]: I1208 14:50:21.872640 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.463340 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.463731 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5f5f4cb9840dfbc9af8e6b09757c681e05506e16a13b823a933445533c35753c"} Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.464992 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.465187 4894 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.466537 4894 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="1b8171043a837f038d258ca7333e2ef0701937ac7b27863cf85f0a3832921686" exitCode=0 Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.466566 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"1b8171043a837f038d258ca7333e2ef0701937ac7b27863cf85f0a3832921686"} Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.466892 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.466927 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.467332 4894 status_manager.go:851] "Failed to get status for pod" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:22 crc kubenswrapper[4894]: E1208 14:50:22.467338 4894 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:22 crc kubenswrapper[4894]: I1208 14:50:22.467642 4894 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.51:6443: connect: connection refused" Dec 08 14:50:23 crc kubenswrapper[4894]: I1208 14:50:23.474609 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9ddf9b469f2779425570c68d25a781120f393ca6304ca9fefe87e85b2f9a1f78"} Dec 08 14:50:23 crc kubenswrapper[4894]: I1208 14:50:23.474659 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6f00c4c49598d5885f16a95494d4a61bf3e7f4536f4982a32c8780e61e4d9d92"} Dec 08 14:50:23 crc kubenswrapper[4894]: I1208 14:50:23.966418 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:50:23 crc kubenswrapper[4894]: I1208 14:50:23.970596 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:50:24 crc kubenswrapper[4894]: I1208 14:50:24.485194 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:24 crc kubenswrapper[4894]: I1208 14:50:24.486971 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:24 crc kubenswrapper[4894]: I1208 14:50:24.485448 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d2dee0e2f00e53a955822277012c4ee234594bfaa433b1111cd2a170b8e18c0c"} Dec 08 14:50:24 crc kubenswrapper[4894]: I1208 14:50:24.487447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fab81a3707971145d13f03c89251be62a2d87b62e5fa84ee3a6ed84708a60f32"} Dec 08 14:50:24 crc kubenswrapper[4894]: I1208 14:50:24.487501 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:50:24 crc kubenswrapper[4894]: I1208 14:50:24.487517 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e7f757fee631ab590ce2b627423b34b4bd6e07f0f1fbee34ccd454b1e388f94f"} Dec 08 14:50:24 crc kubenswrapper[4894]: I1208 14:50:24.487533 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:26 crc kubenswrapper[4894]: I1208 14:50:26.211216 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:26 crc kubenswrapper[4894]: I1208 14:50:26.211693 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:26 crc kubenswrapper[4894]: I1208 14:50:26.216494 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:29 crc kubenswrapper[4894]: I1208 14:50:29.495271 4894 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:29 crc kubenswrapper[4894]: I1208 14:50:29.513325 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:29 crc kubenswrapper[4894]: I1208 14:50:29.513353 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:29 crc kubenswrapper[4894]: I1208 14:50:29.517785 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:29 crc kubenswrapper[4894]: I1208 14:50:29.520183 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b83c6826-d8ef-41fe-8f6d-97fc47e1b64c" Dec 08 14:50:30 crc kubenswrapper[4894]: I1208 14:50:30.517045 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:30 crc kubenswrapper[4894]: I1208 14:50:30.517920 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:31 crc kubenswrapper[4894]: I1208 14:50:31.876525 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 08 14:50:39 crc kubenswrapper[4894]: I1208 14:50:39.066276 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 08 14:50:39 crc kubenswrapper[4894]: I1208 14:50:39.207440 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b83c6826-d8ef-41fe-8f6d-97fc47e1b64c" Dec 08 14:50:39 crc kubenswrapper[4894]: I1208 14:50:39.876352 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 08 14:50:40 crc kubenswrapper[4894]: I1208 14:50:40.261272 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 08 14:50:40 crc kubenswrapper[4894]: I1208 14:50:40.354200 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 08 14:50:40 crc kubenswrapper[4894]: I1208 14:50:40.487332 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 08 14:50:40 crc kubenswrapper[4894]: I1208 14:50:40.910770 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 08 14:50:41 crc kubenswrapper[4894]: I1208 14:50:41.063662 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 08 14:50:41 crc kubenswrapper[4894]: I1208 14:50:41.292954 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 08 14:50:41 crc kubenswrapper[4894]: I1208 14:50:41.324345 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 08 14:50:41 crc kubenswrapper[4894]: I1208 14:50:41.469996 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 08 14:50:41 crc kubenswrapper[4894]: I1208 14:50:41.772321 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 08 14:50:41 crc kubenswrapper[4894]: I1208 14:50:41.975474 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 08 14:50:41 crc kubenswrapper[4894]: I1208 14:50:41.991502 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.073778 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.148174 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.367837 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.467778 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.538106 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.609521 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.622622 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.650591 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.895238 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.961510 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.961987 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 08 14:50:42 crc kubenswrapper[4894]: I1208 14:50:42.983194 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.005877 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.037950 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.048728 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.072300 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.114299 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.203908 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.235217 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.262257 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.308692 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.356114 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.359493 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.510161 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.785019 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.853466 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.860458 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.870386 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.909176 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 08 14:50:43 crc kubenswrapper[4894]: I1208 14:50:43.977987 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.046761 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.094317 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.135540 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.149402 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.170151 4894 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.298715 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.328034 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.348281 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.349089 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.435702 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.448660 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.463435 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.475779 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.586212 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.672331 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.784123 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.912477 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.937043 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 08 14:50:44 crc kubenswrapper[4894]: I1208 14:50:44.942560 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.049624 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.097523 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.126647 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.126890 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.279276 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.303045 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.329508 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.371374 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.536434 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.550867 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.560174 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.568512 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.655921 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.659005 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.804522 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 08 14:50:45 crc kubenswrapper[4894]: I1208 14:50:45.919797 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.018261 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.099190 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.110179 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.111018 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.111367 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.115105 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.115471 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.192576 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.258790 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.338289 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.348168 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.380552 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.427628 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.476576 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.502144 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.546103 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.547202 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.616358 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.643227 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.711370 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.775610 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 08 14:50:46 crc kubenswrapper[4894]: I1208 14:50:46.824510 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.053108 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.347344 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.370444 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.405146 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.453880 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.546834 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.563651 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.565206 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.620556 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.686609 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.694441 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.762330 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.788054 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.861077 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.879894 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 08 14:50:47 crc kubenswrapper[4894]: I1208 14:50:47.920882 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.025106 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.073434 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.123089 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.175289 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.246540 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.332250 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.396470 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.538523 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.575623 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.592027 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.602351 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.629990 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.759916 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.792436 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.793171 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.821029 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.828880 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.864150 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.958506 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.977100 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 08 14:50:48 crc kubenswrapper[4894]: I1208 14:50:48.979100 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.003371 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.012913 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.068657 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.074612 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.110575 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.143501 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.231914 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.248305 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.290128 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.346426 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.360749 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.399004 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.402799 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.466312 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.519016 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.529599 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.555798 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.569499 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.599555 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.645489 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.660325 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.681967 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.692047 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.726831 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.748130 4894 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.797456 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.807678 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.926901 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.977897 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.986621 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 08 14:50:49 crc kubenswrapper[4894]: I1208 14:50:49.987844 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.016388 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.030279 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.099584 4894 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.111647 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.157589 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.179686 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.200152 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.233754 4894 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.261482 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.329666 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.374049 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.436920 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.438100 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.477173 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.540787 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.673666 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.714380 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.766274 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.876288 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 08 14:50:50 crc kubenswrapper[4894]: I1208 14:50:50.999051 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.125983 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.238285 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.242988 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.290121 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.328739 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.363362 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.423507 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.480914 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.547946 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.611214 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.740659 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.842172 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.897012 4894 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.903778 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.904059 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.904568 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.904607 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9112ccff-5eb4-44a5-b333-dac4d0474d73" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.908305 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.922473 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.92245532 podStartE2EDuration="22.92245532s" podCreationTimestamp="2025-12-08 14:50:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:50:51.921585831 +0000 UTC m=+273.021591946" watchObservedRunningTime="2025-12-08 14:50:51.92245532 +0000 UTC m=+273.022461435" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.925784 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.978200 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 08 14:50:51 crc kubenswrapper[4894]: I1208 14:50:51.994276 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.008103 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.053446 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.078444 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.084792 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.148851 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.181447 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.225627 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.289487 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.406423 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.544893 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.678250 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.745939 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.836394 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 08 14:50:52 crc kubenswrapper[4894]: I1208 14:50:52.994460 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.049419 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.049927 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xjtc9"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.050331 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xjtc9" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="registry-server" containerID="cri-o://b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287" gracePeriod=30 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.060166 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t69pb"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.061070 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t69pb" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="registry-server" containerID="cri-o://a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea" gracePeriod=30 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.065958 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25w2m"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.066177 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" podUID="d2838428-72ae-469d-884d-6660de5ce2f1" containerName="marketplace-operator" containerID="cri-o://36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65" gracePeriod=30 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.078781 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4hbv"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.079074 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h4hbv" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="registry-server" containerID="cri-o://6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903" gracePeriod=30 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.087270 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gbbpn"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.087571 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gbbpn" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="registry-server" containerID="cri-o://da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8" gracePeriod=30 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.088197 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.183280 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.324959 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.373362 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.458298 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.470300 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.519124 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.523916 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-utilities\") pod \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.524009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmcbw\" (UniqueName: \"kubernetes.io/projected/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-kube-api-access-bmcbw\") pod \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.526056 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-utilities" (OuterVolumeSpecName: "utilities") pod "fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" (UID: "fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.526546 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.526896 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.530927 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-kube-api-access-bmcbw" (OuterVolumeSpecName: "kube-api-access-bmcbw") pod "0edc032d-9fd8-4fbf-b7c7-9106e78404ee" (UID: "0edc032d-9fd8-4fbf-b7c7-9106e78404ee"). InnerVolumeSpecName "kube-api-access-bmcbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.534885 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-catalog-content\") pod \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.534952 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-catalog-content\") pod \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.534996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-utilities\") pod \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\" (UID: \"0edc032d-9fd8-4fbf-b7c7-9106e78404ee\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.535078 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk84b\" (UniqueName: \"kubernetes.io/projected/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-kube-api-access-wk84b\") pod \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\" (UID: \"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.538348 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-utilities" (OuterVolumeSpecName: "utilities") pod "0edc032d-9fd8-4fbf-b7c7-9106e78404ee" (UID: "0edc032d-9fd8-4fbf-b7c7-9106e78404ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.543924 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-kube-api-access-wk84b" (OuterVolumeSpecName: "kube-api-access-wk84b") pod "fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" (UID: "fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e"). InnerVolumeSpecName "kube-api-access-wk84b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.554456 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.575265 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.605571 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" (UID: "fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.614840 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0edc032d-9fd8-4fbf-b7c7-9106e78404ee" (UID: "0edc032d-9fd8-4fbf-b7c7-9106e78404ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.629143 4894 generic.go:334] "Generic (PLEG): container finished" podID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerID="da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8" exitCode=0 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.629212 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gbbpn" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.629199 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbbpn" event={"ID":"a4d8a8de-7e5f-4387-8a9f-281421db9641","Type":"ContainerDied","Data":"da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.629366 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gbbpn" event={"ID":"a4d8a8de-7e5f-4387-8a9f-281421db9641","Type":"ContainerDied","Data":"20e2e6db693382400d817579908639f3a7affd67972008a0470505f6043715f5"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.629405 4894 scope.go:117] "RemoveContainer" containerID="da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.631251 4894 generic.go:334] "Generic (PLEG): container finished" podID="d2838428-72ae-469d-884d-6660de5ce2f1" containerID="36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65" exitCode=0 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.631272 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.631317 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" event={"ID":"d2838428-72ae-469d-884d-6660de5ce2f1","Type":"ContainerDied","Data":"36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.631341 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-25w2m" event={"ID":"d2838428-72ae-469d-884d-6660de5ce2f1","Type":"ContainerDied","Data":"cedab464b16bca7ff79c3af6419111293487fdab87289ae8efdb2f450e47f0f1"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.633306 4894 generic.go:334] "Generic (PLEG): container finished" podID="ccba7943-00e3-422a-91a6-a85439a4c174" containerID="6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903" exitCode=0 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.633380 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4hbv" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.633399 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4hbv" event={"ID":"ccba7943-00e3-422a-91a6-a85439a4c174","Type":"ContainerDied","Data":"6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.633468 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4hbv" event={"ID":"ccba7943-00e3-422a-91a6-a85439a4c174","Type":"ContainerDied","Data":"d5881c879b060af61c583a68a46d55fac4fc6e188e8fd23414cc621a9dd19bdb"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.635895 4894 generic.go:334] "Generic (PLEG): container finished" podID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerID="b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287" exitCode=0 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636045 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-utilities\") pod \"ccba7943-00e3-422a-91a6-a85439a4c174\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636140 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-catalog-content\") pod \"a4d8a8de-7e5f-4387-8a9f-281421db9641\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636174 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmngr\" (UniqueName: \"kubernetes.io/projected/ccba7943-00e3-422a-91a6-a85439a4c174-kube-api-access-mmngr\") pod \"ccba7943-00e3-422a-91a6-a85439a4c174\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636240 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-utilities\") pod \"a4d8a8de-7e5f-4387-8a9f-281421db9641\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636290 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-catalog-content\") pod \"ccba7943-00e3-422a-91a6-a85439a4c174\" (UID: \"ccba7943-00e3-422a-91a6-a85439a4c174\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636317 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-trusted-ca\") pod \"d2838428-72ae-469d-884d-6660de5ce2f1\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636342 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xhcj\" (UniqueName: \"kubernetes.io/projected/d2838428-72ae-469d-884d-6660de5ce2f1-kube-api-access-2xhcj\") pod \"d2838428-72ae-469d-884d-6660de5ce2f1\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636367 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-operator-metrics\") pod \"d2838428-72ae-469d-884d-6660de5ce2f1\" (UID: \"d2838428-72ae-469d-884d-6660de5ce2f1\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636412 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnztk\" (UniqueName: \"kubernetes.io/projected/a4d8a8de-7e5f-4387-8a9f-281421db9641-kube-api-access-tnztk\") pod \"a4d8a8de-7e5f-4387-8a9f-281421db9641\" (UID: \"a4d8a8de-7e5f-4387-8a9f-281421db9641\") " Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636588 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xjtc9" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636794 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-utilities" (OuterVolumeSpecName: "utilities") pod "ccba7943-00e3-422a-91a6-a85439a4c174" (UID: "ccba7943-00e3-422a-91a6-a85439a4c174"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.637031 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-utilities" (OuterVolumeSpecName: "utilities") pod "a4d8a8de-7e5f-4387-8a9f-281421db9641" (UID: "a4d8a8de-7e5f-4387-8a9f-281421db9641"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636049 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjtc9" event={"ID":"0edc032d-9fd8-4fbf-b7c7-9106e78404ee","Type":"ContainerDied","Data":"b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.637093 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xjtc9" event={"ID":"0edc032d-9fd8-4fbf-b7c7-9106e78404ee","Type":"ContainerDied","Data":"7ea177201e1e0a3630d94ff8c8d464410613c613aa7cda98e45d8fa2acb84ecd"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.636659 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmcbw\" (UniqueName: \"kubernetes.io/projected/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-kube-api-access-bmcbw\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.637510 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.637530 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.637543 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edc032d-9fd8-4fbf-b7c7-9106e78404ee-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.637557 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk84b\" (UniqueName: \"kubernetes.io/projected/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-kube-api-access-wk84b\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.637594 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.638765 4894 generic.go:334] "Generic (PLEG): container finished" podID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerID="a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea" exitCode=0 Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.638798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t69pb" event={"ID":"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e","Type":"ContainerDied","Data":"a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.638831 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t69pb" event={"ID":"fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e","Type":"ContainerDied","Data":"913bd1c1fb3b0da6a4cccc41e2cb9e0b2ce96e29d538113c96c734cde7813515"} Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.638899 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t69pb" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.639631 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d8a8de-7e5f-4387-8a9f-281421db9641-kube-api-access-tnztk" (OuterVolumeSpecName: "kube-api-access-tnztk") pod "a4d8a8de-7e5f-4387-8a9f-281421db9641" (UID: "a4d8a8de-7e5f-4387-8a9f-281421db9641"). InnerVolumeSpecName "kube-api-access-tnztk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.639847 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2838428-72ae-469d-884d-6660de5ce2f1-kube-api-access-2xhcj" (OuterVolumeSpecName: "kube-api-access-2xhcj") pod "d2838428-72ae-469d-884d-6660de5ce2f1" (UID: "d2838428-72ae-469d-884d-6660de5ce2f1"). InnerVolumeSpecName "kube-api-access-2xhcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.644130 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccba7943-00e3-422a-91a6-a85439a4c174-kube-api-access-mmngr" (OuterVolumeSpecName: "kube-api-access-mmngr") pod "ccba7943-00e3-422a-91a6-a85439a4c174" (UID: "ccba7943-00e3-422a-91a6-a85439a4c174"). InnerVolumeSpecName "kube-api-access-mmngr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.646295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d2838428-72ae-469d-884d-6660de5ce2f1" (UID: "d2838428-72ae-469d-884d-6660de5ce2f1"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.646736 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d2838428-72ae-469d-884d-6660de5ce2f1" (UID: "d2838428-72ae-469d-884d-6660de5ce2f1"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.652859 4894 scope.go:117] "RemoveContainer" containerID="cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.657247 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccba7943-00e3-422a-91a6-a85439a4c174" (UID: "ccba7943-00e3-422a-91a6-a85439a4c174"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.671201 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t69pb"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.673874 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t69pb"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.684193 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.689913 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xjtc9"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.693285 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xjtc9"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.697624 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.699683 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.703842 4894 scope.go:117] "RemoveContainer" containerID="dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.716645 4894 scope.go:117] "RemoveContainer" containerID="da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.717573 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8\": container with ID starting with da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8 not found: ID does not exist" containerID="da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.717609 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8"} err="failed to get container status \"da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8\": rpc error: code = NotFound desc = could not find container \"da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8\": container with ID starting with da4268dc04fdaca5e5edd6f4846a0d0f6e1e20f7278f59c23e9e79e880bd42b8 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.717638 4894 scope.go:117] "RemoveContainer" containerID="cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.720979 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab\": container with ID starting with cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab not found: ID does not exist" containerID="cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.721030 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab"} err="failed to get container status \"cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab\": rpc error: code = NotFound desc = could not find container \"cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab\": container with ID starting with cd8ac5c14c698800c1cb9868f17dc6dfe30a2201d55951426314b829b84894ab not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.721058 4894 scope.go:117] "RemoveContainer" containerID="dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.721364 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646\": container with ID starting with dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646 not found: ID does not exist" containerID="dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.721396 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646"} err="failed to get container status \"dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646\": rpc error: code = NotFound desc = could not find container \"dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646\": container with ID starting with dfa33df626e9f65a0f31cdfa6d917b08f0fecc557727b0d2a620b72b981f3646 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.721420 4894 scope.go:117] "RemoveContainer" containerID="36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739389 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739435 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmngr\" (UniqueName: \"kubernetes.io/projected/ccba7943-00e3-422a-91a6-a85439a4c174-kube-api-access-mmngr\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739449 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739461 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccba7943-00e3-422a-91a6-a85439a4c174-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739479 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739490 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xhcj\" (UniqueName: \"kubernetes.io/projected/d2838428-72ae-469d-884d-6660de5ce2f1-kube-api-access-2xhcj\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739503 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d2838428-72ae-469d-884d-6660de5ce2f1-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.739516 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnztk\" (UniqueName: \"kubernetes.io/projected/a4d8a8de-7e5f-4387-8a9f-281421db9641-kube-api-access-tnztk\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.743313 4894 scope.go:117] "RemoveContainer" containerID="36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.743690 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65\": container with ID starting with 36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65 not found: ID does not exist" containerID="36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.743718 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65"} err="failed to get container status \"36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65\": rpc error: code = NotFound desc = could not find container \"36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65\": container with ID starting with 36f4955b3293a50f5917b00cf5935dcfd7fbc8941c3849a9d0849be8b137ab65 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.743743 4894 scope.go:117] "RemoveContainer" containerID="6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.757526 4894 scope.go:117] "RemoveContainer" containerID="dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.768868 4894 scope.go:117] "RemoveContainer" containerID="5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.770576 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.788162 4894 scope.go:117] "RemoveContainer" containerID="6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.788667 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903\": container with ID starting with 6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903 not found: ID does not exist" containerID="6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.788722 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903"} err="failed to get container status \"6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903\": rpc error: code = NotFound desc = could not find container \"6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903\": container with ID starting with 6257069328b4b34db7f62f0ef260d7742f04d04d8b4d567f574640f3ab16a903 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.788762 4894 scope.go:117] "RemoveContainer" containerID="dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.789116 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343\": container with ID starting with dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343 not found: ID does not exist" containerID="dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.789148 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343"} err="failed to get container status \"dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343\": rpc error: code = NotFound desc = could not find container \"dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343\": container with ID starting with dddb303215ef16d8b9935d307942122c306754a2aba18a2e54d6c8ffa2bb0343 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.789174 4894 scope.go:117] "RemoveContainer" containerID="5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.789755 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021\": container with ID starting with 5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021 not found: ID does not exist" containerID="5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.789803 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021"} err="failed to get container status \"5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021\": rpc error: code = NotFound desc = could not find container \"5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021\": container with ID starting with 5a17e345f0dcb4c88cf4ff6bb223f2ede8ac8280f11f6665695498d50f319021 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.789949 4894 scope.go:117] "RemoveContainer" containerID="b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.793155 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4d8a8de-7e5f-4387-8a9f-281421db9641" (UID: "a4d8a8de-7e5f-4387-8a9f-281421db9641"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.808181 4894 scope.go:117] "RemoveContainer" containerID="e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.840949 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4d8a8de-7e5f-4387-8a9f-281421db9641-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.867767 4894 scope.go:117] "RemoveContainer" containerID="1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.880969 4894 scope.go:117] "RemoveContainer" containerID="b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.881384 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287\": container with ID starting with b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287 not found: ID does not exist" containerID="b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.881510 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287"} err="failed to get container status \"b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287\": rpc error: code = NotFound desc = could not find container \"b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287\": container with ID starting with b6db49a5eba30e45f863d478a36d47547683d2d76fb49668ffce2aa79e0ca287 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.881606 4894 scope.go:117] "RemoveContainer" containerID="e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.882280 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca\": container with ID starting with e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca not found: ID does not exist" containerID="e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.882310 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca"} err="failed to get container status \"e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca\": rpc error: code = NotFound desc = could not find container \"e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca\": container with ID starting with e4d3baed3b47773dc2fd83e49cd277025b3266f4d0565b634cd5a9fceff58fca not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.882330 4894 scope.go:117] "RemoveContainer" containerID="1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.882646 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519\": container with ID starting with 1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519 not found: ID does not exist" containerID="1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.882743 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519"} err="failed to get container status \"1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519\": rpc error: code = NotFound desc = could not find container \"1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519\": container with ID starting with 1180803979bb295434d70bc80b64f414b9c9a557dc93b50a4cd277eaa311b519 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.882812 4894 scope.go:117] "RemoveContainer" containerID="a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.895159 4894 scope.go:117] "RemoveContainer" containerID="95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.912778 4894 scope.go:117] "RemoveContainer" containerID="6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.922634 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.931551 4894 scope.go:117] "RemoveContainer" containerID="a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.932094 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea\": container with ID starting with a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea not found: ID does not exist" containerID="a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.932138 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea"} err="failed to get container status \"a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea\": rpc error: code = NotFound desc = could not find container \"a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea\": container with ID starting with a7f8407bf52122386676fc1368c9d2b0541314e18176680e0ef1d9ef3a0782ea not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.932165 4894 scope.go:117] "RemoveContainer" containerID="95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.932512 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914\": container with ID starting with 95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914 not found: ID does not exist" containerID="95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.932555 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914"} err="failed to get container status \"95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914\": rpc error: code = NotFound desc = could not find container \"95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914\": container with ID starting with 95c821de7a8643116e1a4f83868ce774a9cff5afbfe0ab0a566f13a6794ca914 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.932588 4894 scope.go:117] "RemoveContainer" containerID="6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04" Dec 08 14:50:53 crc kubenswrapper[4894]: E1208 14:50:53.933008 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04\": container with ID starting with 6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04 not found: ID does not exist" containerID="6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.933032 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04"} err="failed to get container status \"6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04\": rpc error: code = NotFound desc = could not find container \"6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04\": container with ID starting with 6a85b8356605e2c8f40b1f21d5bd51a959cb752e915f6ffc065f5bef13550c04 not found: ID does not exist" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.961692 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.966548 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25w2m"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.974543 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-25w2m"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.980789 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gbbpn"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.984647 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gbbpn"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.987579 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4hbv"] Dec 08 14:50:53 crc kubenswrapper[4894]: I1208 14:50:53.990661 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4hbv"] Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.064044 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.115680 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.308236 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.438273 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450228 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7db"] Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450420 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450436 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450448 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450456 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450468 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450474 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450485 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450491 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450499 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450505 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450512 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450518 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450524 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450530 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450539 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450550 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450562 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450569 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450577 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450583 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450591 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450597 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="extract-utilities" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450604 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450610 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="extract-content" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450617 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" containerName="installer" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450623 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" containerName="installer" Dec 08 14:50:54 crc kubenswrapper[4894]: E1208 14:50:54.450637 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2838428-72ae-469d-884d-6660de5ce2f1" containerName="marketplace-operator" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450648 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2838428-72ae-469d-884d-6660de5ce2f1" containerName="marketplace-operator" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450760 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450770 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450779 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2838428-72ae-469d-884d-6660de5ce2f1" containerName="marketplace-operator" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450792 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450800 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" containerName="registry-server" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.450806 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="40041abc-c8e0-46d9-acdc-c4789f020b6c" containerName="installer" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.451212 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.453319 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.453584 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.453783 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.453969 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.454436 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.459322 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7db"] Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.460970 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.548955 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbzmj\" (UniqueName: \"kubernetes.io/projected/e1fa33df-7fab-40bc-8a30-9004002f770a-kube-api-access-zbzmj\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.549026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1fa33df-7fab-40bc-8a30-9004002f770a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.549066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e1fa33df-7fab-40bc-8a30-9004002f770a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.613249 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.649731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1fa33df-7fab-40bc-8a30-9004002f770a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.649804 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e1fa33df-7fab-40bc-8a30-9004002f770a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.649872 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbzmj\" (UniqueName: \"kubernetes.io/projected/e1fa33df-7fab-40bc-8a30-9004002f770a-kube-api-access-zbzmj\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.651124 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1fa33df-7fab-40bc-8a30-9004002f770a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.655626 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e1fa33df-7fab-40bc-8a30-9004002f770a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.675035 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbzmj\" (UniqueName: \"kubernetes.io/projected/e1fa33df-7fab-40bc-8a30-9004002f770a-kube-api-access-zbzmj\") pod \"marketplace-operator-79b997595-bl7db\" (UID: \"e1fa33df-7fab-40bc-8a30-9004002f770a\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.760434 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.770681 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:54 crc kubenswrapper[4894]: I1208 14:50:54.941248 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.019671 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.023262 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.233861 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0edc032d-9fd8-4fbf-b7c7-9106e78404ee" path="/var/lib/kubelet/pods/0edc032d-9fd8-4fbf-b7c7-9106e78404ee/volumes" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.234787 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d8a8de-7e5f-4387-8a9f-281421db9641" path="/var/lib/kubelet/pods/a4d8a8de-7e5f-4387-8a9f-281421db9641/volumes" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.235470 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccba7943-00e3-422a-91a6-a85439a4c174" path="/var/lib/kubelet/pods/ccba7943-00e3-422a-91a6-a85439a4c174/volumes" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.236789 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2838428-72ae-469d-884d-6660de5ce2f1" path="/var/lib/kubelet/pods/d2838428-72ae-469d-884d-6660de5ce2f1/volumes" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.237391 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e" path="/var/lib/kubelet/pods/fcf5ad5c-9ffd-451f-a433-6dc0ba71bf1e/volumes" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.238088 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7db"] Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.302935 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.393897 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.399560 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.664246 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" event={"ID":"e1fa33df-7fab-40bc-8a30-9004002f770a","Type":"ContainerStarted","Data":"0c018c6aae5662b1ddae6901522044eca594e0bac026d3d64532ae7954c78e46"} Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.664299 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" event={"ID":"e1fa33df-7fab-40bc-8a30-9004002f770a","Type":"ContainerStarted","Data":"07e508811ecf21453d160631205c0b159397aaf362670bb8f3663805e426ef4e"} Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.666208 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.668680 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.701951 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bl7db" podStartSLOduration=2.7019267510000002 podStartE2EDuration="2.701926751s" podCreationTimestamp="2025-12-08 14:50:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:50:55.684433608 +0000 UTC m=+276.784439723" watchObservedRunningTime="2025-12-08 14:50:55.701926751 +0000 UTC m=+276.801932866" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.831652 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 08 14:50:55 crc kubenswrapper[4894]: I1208 14:50:55.861552 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 08 14:50:56 crc kubenswrapper[4894]: I1208 14:50:56.155112 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 08 14:50:56 crc kubenswrapper[4894]: I1208 14:50:56.382372 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 08 14:50:57 crc kubenswrapper[4894]: I1208 14:50:57.123795 4894 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 08 14:50:57 crc kubenswrapper[4894]: I1208 14:50:57.657880 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 08 14:50:57 crc kubenswrapper[4894]: I1208 14:50:57.909347 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 08 14:50:58 crc kubenswrapper[4894]: I1208 14:50:58.076566 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 08 14:51:03 crc kubenswrapper[4894]: I1208 14:51:03.206516 4894 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 08 14:51:03 crc kubenswrapper[4894]: I1208 14:51:03.207395 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://12a7eda32e24169a241711690e677a618a6f25a64891c7128bfc953dcce265e9" gracePeriod=5 Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.731139 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.731887 4894 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="12a7eda32e24169a241711690e677a618a6f25a64891c7128bfc953dcce265e9" exitCode=137 Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.767941 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.768408 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907382 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907437 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907476 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907491 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907516 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907640 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907748 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907740 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.907770 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:51:08 crc kubenswrapper[4894]: I1208 14:51:08.915980 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.009236 4894 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.009687 4894 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.009885 4894 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.010034 4894 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.010164 4894 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.203436 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.738395 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.738487 4894 scope.go:117] "RemoveContainer" containerID="12a7eda32e24169a241711690e677a618a6f25a64891c7128bfc953dcce265e9" Dec 08 14:51:09 crc kubenswrapper[4894]: I1208 14:51:09.738593 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 08 14:51:18 crc kubenswrapper[4894]: I1208 14:51:18.953860 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-52vvt"] Dec 08 14:51:18 crc kubenswrapper[4894]: E1208 14:51:18.954807 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 08 14:51:18 crc kubenswrapper[4894]: I1208 14:51:18.954887 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 08 14:51:18 crc kubenswrapper[4894]: I1208 14:51:18.955142 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 08 14:51:18 crc kubenswrapper[4894]: I1208 14:51:18.956275 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:18 crc kubenswrapper[4894]: I1208 14:51:18.960675 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52vvt"] Dec 08 14:51:18 crc kubenswrapper[4894]: I1208 14:51:18.962563 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.054651 4894 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.129056 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-utilities\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.129135 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-catalog-content\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.129157 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vssps\" (UniqueName: \"kubernetes.io/projected/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-kube-api-access-vssps\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.148747 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qc7vh"] Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.149742 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.154219 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.161196 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qc7vh"] Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.230612 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-utilities\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.230681 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-catalog-content\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.230711 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vssps\" (UniqueName: \"kubernetes.io/projected/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-kube-api-access-vssps\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.230756 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxkrb\" (UniqueName: \"kubernetes.io/projected/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-kube-api-access-pxkrb\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.230776 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-utilities\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.230791 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-catalog-content\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.231113 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-utilities\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.231152 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-catalog-content\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.249401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vssps\" (UniqueName: \"kubernetes.io/projected/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-kube-api-access-vssps\") pod \"redhat-operators-52vvt\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.276132 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.284709 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.332332 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxkrb\" (UniqueName: \"kubernetes.io/projected/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-kube-api-access-pxkrb\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.332393 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-utilities\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.332414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-catalog-content\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.333087 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-catalog-content\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.333469 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-utilities\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.356585 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxkrb\" (UniqueName: \"kubernetes.io/projected/c0e4e418-79da-43c5-8d59-1d1f6a7e2323-kube-api-access-pxkrb\") pod \"redhat-marketplace-qc7vh\" (UID: \"c0e4e418-79da-43c5-8d59-1d1f6a7e2323\") " pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.462696 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.518237 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-52vvt"] Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.644788 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qc7vh"] Dec 08 14:51:19 crc kubenswrapper[4894]: W1208 14:51:19.646996 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e4e418_79da_43c5_8d59_1d1f6a7e2323.slice/crio-4176be80f022a535f3a08058a3b890ad0c11030696fac319eb244049f3def413 WatchSource:0}: Error finding container 4176be80f022a535f3a08058a3b890ad0c11030696fac319eb244049f3def413: Status 404 returned error can't find the container with id 4176be80f022a535f3a08058a3b890ad0c11030696fac319eb244049f3def413 Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.812628 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qc7vh" event={"ID":"c0e4e418-79da-43c5-8d59-1d1f6a7e2323","Type":"ContainerStarted","Data":"4176be80f022a535f3a08058a3b890ad0c11030696fac319eb244049f3def413"} Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.814101 4894 generic.go:334] "Generic (PLEG): container finished" podID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerID="0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9" exitCode=0 Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.814144 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vvt" event={"ID":"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3","Type":"ContainerDied","Data":"0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9"} Dec 08 14:51:19 crc kubenswrapper[4894]: I1208 14:51:19.814182 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vvt" event={"ID":"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3","Type":"ContainerStarted","Data":"3253988fdffbecf04ade87ec75761a2668d3582a6fa839d85745dbe95e6185bc"} Dec 08 14:51:20 crc kubenswrapper[4894]: I1208 14:51:20.820837 4894 generic.go:334] "Generic (PLEG): container finished" podID="c0e4e418-79da-43c5-8d59-1d1f6a7e2323" containerID="0e99ed2e6bb2c35cdf48dd1d5e00d1047295f382934df5723c299f2d3b42d6c6" exitCode=0 Dec 08 14:51:20 crc kubenswrapper[4894]: I1208 14:51:20.820950 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qc7vh" event={"ID":"c0e4e418-79da-43c5-8d59-1d1f6a7e2323","Type":"ContainerDied","Data":"0e99ed2e6bb2c35cdf48dd1d5e00d1047295f382934df5723c299f2d3b42d6c6"} Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.348398 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bhxcd"] Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.349587 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.351236 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.361758 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhxcd"] Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.458898 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a7ac74-25ab-4f68-8aee-adbbf5defd99-catalog-content\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.459316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a7ac74-25ab-4f68-8aee-adbbf5defd99-utilities\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.459482 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjwjs\" (UniqueName: \"kubernetes.io/projected/28a7ac74-25ab-4f68-8aee-adbbf5defd99-kube-api-access-pjwjs\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.549056 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s7cr5"] Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.550066 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.551764 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.560733 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7cr5"] Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.561019 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdpcz\" (UniqueName: \"kubernetes.io/projected/9393c174-85f4-4267-80a1-e7c1075104b5-kube-api-access-mdpcz\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.561194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjwjs\" (UniqueName: \"kubernetes.io/projected/28a7ac74-25ab-4f68-8aee-adbbf5defd99-kube-api-access-pjwjs\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.561323 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a7ac74-25ab-4f68-8aee-adbbf5defd99-catalog-content\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.561467 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-utilities\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.561551 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a7ac74-25ab-4f68-8aee-adbbf5defd99-utilities\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.561640 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-catalog-content\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.562172 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28a7ac74-25ab-4f68-8aee-adbbf5defd99-utilities\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.562324 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28a7ac74-25ab-4f68-8aee-adbbf5defd99-catalog-content\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.585681 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjwjs\" (UniqueName: \"kubernetes.io/projected/28a7ac74-25ab-4f68-8aee-adbbf5defd99-kube-api-access-pjwjs\") pod \"certified-operators-bhxcd\" (UID: \"28a7ac74-25ab-4f68-8aee-adbbf5defd99\") " pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.662449 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdpcz\" (UniqueName: \"kubernetes.io/projected/9393c174-85f4-4267-80a1-e7c1075104b5-kube-api-access-mdpcz\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.662522 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-utilities\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.662551 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-catalog-content\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.663127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-catalog-content\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.663163 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-utilities\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.665393 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.682433 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdpcz\" (UniqueName: \"kubernetes.io/projected/9393c174-85f4-4267-80a1-e7c1075104b5-kube-api-access-mdpcz\") pod \"community-operators-s7cr5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.827888 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qc7vh" event={"ID":"c0e4e418-79da-43c5-8d59-1d1f6a7e2323","Type":"ContainerStarted","Data":"82c38cd16d2f52e2ece0a20c0d0f5deef5b0b9bea441eeda5ba438c7adf7d1e8"} Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.831377 4894 generic.go:334] "Generic (PLEG): container finished" podID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerID="f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474" exitCode=0 Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.831425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vvt" event={"ID":"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3","Type":"ContainerDied","Data":"f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474"} Dec 08 14:51:21 crc kubenswrapper[4894]: I1208 14:51:21.869147 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.049501 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7cr5"] Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.053026 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bhxcd"] Dec 08 14:51:22 crc kubenswrapper[4894]: W1208 14:51:22.056892 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28a7ac74_25ab_4f68_8aee_adbbf5defd99.slice/crio-ae12459ae694ff31fa0b6e57bbc89eb1929ab29c47b4693fcb1842fdcc52406d WatchSource:0}: Error finding container ae12459ae694ff31fa0b6e57bbc89eb1929ab29c47b4693fcb1842fdcc52406d: Status 404 returned error can't find the container with id ae12459ae694ff31fa0b6e57bbc89eb1929ab29c47b4693fcb1842fdcc52406d Dec 08 14:51:22 crc kubenswrapper[4894]: W1208 14:51:22.058355 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9393c174_85f4_4267_80a1_e7c1075104b5.slice/crio-8fffc084b6575d5e449ccc63e824b1eb0a0f8f9fafdd1ec495226bded68ca452 WatchSource:0}: Error finding container 8fffc084b6575d5e449ccc63e824b1eb0a0f8f9fafdd1ec495226bded68ca452: Status 404 returned error can't find the container with id 8fffc084b6575d5e449ccc63e824b1eb0a0f8f9fafdd1ec495226bded68ca452 Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.839529 4894 generic.go:334] "Generic (PLEG): container finished" podID="28a7ac74-25ab-4f68-8aee-adbbf5defd99" containerID="e4a9434d6c9de0f9bf77f54583df198f3c1c4a8f01e48123a24a0299ca676023" exitCode=0 Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.839629 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhxcd" event={"ID":"28a7ac74-25ab-4f68-8aee-adbbf5defd99","Type":"ContainerDied","Data":"e4a9434d6c9de0f9bf77f54583df198f3c1c4a8f01e48123a24a0299ca676023"} Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.839917 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhxcd" event={"ID":"28a7ac74-25ab-4f68-8aee-adbbf5defd99","Type":"ContainerStarted","Data":"ae12459ae694ff31fa0b6e57bbc89eb1929ab29c47b4693fcb1842fdcc52406d"} Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.845971 4894 generic.go:334] "Generic (PLEG): container finished" podID="c0e4e418-79da-43c5-8d59-1d1f6a7e2323" containerID="82c38cd16d2f52e2ece0a20c0d0f5deef5b0b9bea441eeda5ba438c7adf7d1e8" exitCode=0 Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.846038 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qc7vh" event={"ID":"c0e4e418-79da-43c5-8d59-1d1f6a7e2323","Type":"ContainerDied","Data":"82c38cd16d2f52e2ece0a20c0d0f5deef5b0b9bea441eeda5ba438c7adf7d1e8"} Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.850830 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vvt" event={"ID":"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3","Type":"ContainerStarted","Data":"0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d"} Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.852722 4894 generic.go:334] "Generic (PLEG): container finished" podID="9393c174-85f4-4267-80a1-e7c1075104b5" containerID="1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7" exitCode=0 Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.852759 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7cr5" event={"ID":"9393c174-85f4-4267-80a1-e7c1075104b5","Type":"ContainerDied","Data":"1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7"} Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.852776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7cr5" event={"ID":"9393c174-85f4-4267-80a1-e7c1075104b5","Type":"ContainerStarted","Data":"8fffc084b6575d5e449ccc63e824b1eb0a0f8f9fafdd1ec495226bded68ca452"} Dec 08 14:51:22 crc kubenswrapper[4894]: I1208 14:51:22.871005 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-52vvt" podStartSLOduration=2.488491257 podStartE2EDuration="4.870987144s" podCreationTimestamp="2025-12-08 14:51:18 +0000 UTC" firstStartedPulling="2025-12-08 14:51:19.815284243 +0000 UTC m=+300.915290358" lastFinishedPulling="2025-12-08 14:51:22.19778013 +0000 UTC m=+303.297786245" observedRunningTime="2025-12-08 14:51:22.870133707 +0000 UTC m=+303.970139822" watchObservedRunningTime="2025-12-08 14:51:22.870987144 +0000 UTC m=+303.970993259" Dec 08 14:51:23 crc kubenswrapper[4894]: I1208 14:51:23.859767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qc7vh" event={"ID":"c0e4e418-79da-43c5-8d59-1d1f6a7e2323","Type":"ContainerStarted","Data":"f064d9d72af1a8bd4e6990db8db81f6229c7bfea3e257c80e3eaf9a8c57acca8"} Dec 08 14:51:23 crc kubenswrapper[4894]: I1208 14:51:23.862403 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7cr5" event={"ID":"9393c174-85f4-4267-80a1-e7c1075104b5","Type":"ContainerStarted","Data":"0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae"} Dec 08 14:51:23 crc kubenswrapper[4894]: I1208 14:51:23.865300 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhxcd" event={"ID":"28a7ac74-25ab-4f68-8aee-adbbf5defd99","Type":"ContainerStarted","Data":"dd377e50e2d5617c80bb3432db47d57198bff6d79d5cef453edfd51de9b74b45"} Dec 08 14:51:23 crc kubenswrapper[4894]: I1208 14:51:23.884301 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qc7vh" podStartSLOduration=2.268228743 podStartE2EDuration="4.884283573s" podCreationTimestamp="2025-12-08 14:51:19 +0000 UTC" firstStartedPulling="2025-12-08 14:51:20.823080448 +0000 UTC m=+301.923086563" lastFinishedPulling="2025-12-08 14:51:23.439135268 +0000 UTC m=+304.539141393" observedRunningTime="2025-12-08 14:51:23.881025755 +0000 UTC m=+304.981031870" watchObservedRunningTime="2025-12-08 14:51:23.884283573 +0000 UTC m=+304.984289688" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.393585 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xgp2c"] Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.393874 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" podUID="0dff8795-f189-46d1-8d8b-aae4dad68c63" containerName="controller-manager" containerID="cri-o://653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae" gracePeriod=30 Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.492633 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt"] Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.738936 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.806702 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dff8795-f189-46d1-8d8b-aae4dad68c63-serving-cert\") pod \"0dff8795-f189-46d1-8d8b-aae4dad68c63\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.806795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-proxy-ca-bundles\") pod \"0dff8795-f189-46d1-8d8b-aae4dad68c63\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.806860 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-config\") pod \"0dff8795-f189-46d1-8d8b-aae4dad68c63\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.806899 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-client-ca\") pod \"0dff8795-f189-46d1-8d8b-aae4dad68c63\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.806922 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2kkb\" (UniqueName: \"kubernetes.io/projected/0dff8795-f189-46d1-8d8b-aae4dad68c63-kube-api-access-m2kkb\") pod \"0dff8795-f189-46d1-8d8b-aae4dad68c63\" (UID: \"0dff8795-f189-46d1-8d8b-aae4dad68c63\") " Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.807490 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-client-ca" (OuterVolumeSpecName: "client-ca") pod "0dff8795-f189-46d1-8d8b-aae4dad68c63" (UID: "0dff8795-f189-46d1-8d8b-aae4dad68c63"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.807651 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-config" (OuterVolumeSpecName: "config") pod "0dff8795-f189-46d1-8d8b-aae4dad68c63" (UID: "0dff8795-f189-46d1-8d8b-aae4dad68c63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.807995 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0dff8795-f189-46d1-8d8b-aae4dad68c63" (UID: "0dff8795-f189-46d1-8d8b-aae4dad68c63"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.812656 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dff8795-f189-46d1-8d8b-aae4dad68c63-kube-api-access-m2kkb" (OuterVolumeSpecName: "kube-api-access-m2kkb") pod "0dff8795-f189-46d1-8d8b-aae4dad68c63" (UID: "0dff8795-f189-46d1-8d8b-aae4dad68c63"). InnerVolumeSpecName "kube-api-access-m2kkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.812791 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dff8795-f189-46d1-8d8b-aae4dad68c63-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0dff8795-f189-46d1-8d8b-aae4dad68c63" (UID: "0dff8795-f189-46d1-8d8b-aae4dad68c63"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.872017 4894 generic.go:334] "Generic (PLEG): container finished" podID="9393c174-85f4-4267-80a1-e7c1075104b5" containerID="0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae" exitCode=0 Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.872102 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7cr5" event={"ID":"9393c174-85f4-4267-80a1-e7c1075104b5","Type":"ContainerDied","Data":"0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae"} Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.874548 4894 generic.go:334] "Generic (PLEG): container finished" podID="0dff8795-f189-46d1-8d8b-aae4dad68c63" containerID="653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae" exitCode=0 Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.874601 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.874653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" event={"ID":"0dff8795-f189-46d1-8d8b-aae4dad68c63","Type":"ContainerDied","Data":"653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae"} Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.874707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xgp2c" event={"ID":"0dff8795-f189-46d1-8d8b-aae4dad68c63","Type":"ContainerDied","Data":"0014edab7552393d4e55a271efbdff9de8220b76d80436f3ec2a550db3a58838"} Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.874727 4894 scope.go:117] "RemoveContainer" containerID="653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.885761 4894 generic.go:334] "Generic (PLEG): container finished" podID="28a7ac74-25ab-4f68-8aee-adbbf5defd99" containerID="dd377e50e2d5617c80bb3432db47d57198bff6d79d5cef453edfd51de9b74b45" exitCode=0 Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.885840 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhxcd" event={"ID":"28a7ac74-25ab-4f68-8aee-adbbf5defd99","Type":"ContainerDied","Data":"dd377e50e2d5617c80bb3432db47d57198bff6d79d5cef453edfd51de9b74b45"} Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.886061 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" podUID="552313d0-57ae-4e50-ae23-d7a568e1afcd" containerName="route-controller-manager" containerID="cri-o://e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884" gracePeriod=30 Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.907910 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.907944 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.907955 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0dff8795-f189-46d1-8d8b-aae4dad68c63-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.907964 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2kkb\" (UniqueName: \"kubernetes.io/projected/0dff8795-f189-46d1-8d8b-aae4dad68c63-kube-api-access-m2kkb\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.907975 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0dff8795-f189-46d1-8d8b-aae4dad68c63-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.924431 4894 scope.go:117] "RemoveContainer" containerID="653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae" Dec 08 14:51:24 crc kubenswrapper[4894]: E1208 14:51:24.924786 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae\": container with ID starting with 653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae not found: ID does not exist" containerID="653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.924940 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae"} err="failed to get container status \"653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae\": rpc error: code = NotFound desc = could not find container \"653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae\": container with ID starting with 653713d9ddd22eb1428ee5636df505bfe81d264d8fa35f1a47fc16fb958e26ae not found: ID does not exist" Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.944922 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xgp2c"] Dec 08 14:51:24 crc kubenswrapper[4894]: I1208 14:51:24.948085 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xgp2c"] Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.078302 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f54f59f4f-dkkps"] Dec 08 14:51:25 crc kubenswrapper[4894]: E1208 14:51:25.078652 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dff8795-f189-46d1-8d8b-aae4dad68c63" containerName="controller-manager" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.078685 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dff8795-f189-46d1-8d8b-aae4dad68c63" containerName="controller-manager" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.078878 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dff8795-f189-46d1-8d8b-aae4dad68c63" containerName="controller-manager" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.079441 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.081717 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.082272 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.083326 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.083923 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.083987 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.088882 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.092887 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.102060 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f54f59f4f-dkkps"] Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.110253 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-serving-cert\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.110328 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-config\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.110372 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-client-ca\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.110416 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-proxy-ca-bundles\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.110486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngb2v\" (UniqueName: \"kubernetes.io/projected/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-kube-api-access-ngb2v\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.207040 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dff8795-f189-46d1-8d8b-aae4dad68c63" path="/var/lib/kubelet/pods/0dff8795-f189-46d1-8d8b-aae4dad68c63/volumes" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.211731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-config\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.211793 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-client-ca\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.211854 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-proxy-ca-bundles\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.211878 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngb2v\" (UniqueName: \"kubernetes.io/projected/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-kube-api-access-ngb2v\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.211903 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-serving-cert\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.213650 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-config\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.213694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-client-ca\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.214505 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-proxy-ca-bundles\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.216491 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-serving-cert\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.231399 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngb2v\" (UniqueName: \"kubernetes.io/projected/c42ba707-9381-4fc9-884d-9cbe5ac31cb9-kube-api-access-ngb2v\") pod \"controller-manager-7f54f59f4f-dkkps\" (UID: \"c42ba707-9381-4fc9-884d-9cbe5ac31cb9\") " pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.272891 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.313390 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-config\") pod \"552313d0-57ae-4e50-ae23-d7a568e1afcd\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.313494 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552313d0-57ae-4e50-ae23-d7a568e1afcd-serving-cert\") pod \"552313d0-57ae-4e50-ae23-d7a568e1afcd\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.313541 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-client-ca\") pod \"552313d0-57ae-4e50-ae23-d7a568e1afcd\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.313578 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvn5v\" (UniqueName: \"kubernetes.io/projected/552313d0-57ae-4e50-ae23-d7a568e1afcd-kube-api-access-dvn5v\") pod \"552313d0-57ae-4e50-ae23-d7a568e1afcd\" (UID: \"552313d0-57ae-4e50-ae23-d7a568e1afcd\") " Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.314355 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-client-ca" (OuterVolumeSpecName: "client-ca") pod "552313d0-57ae-4e50-ae23-d7a568e1afcd" (UID: "552313d0-57ae-4e50-ae23-d7a568e1afcd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.314376 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-config" (OuterVolumeSpecName: "config") pod "552313d0-57ae-4e50-ae23-d7a568e1afcd" (UID: "552313d0-57ae-4e50-ae23-d7a568e1afcd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.318640 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/552313d0-57ae-4e50-ae23-d7a568e1afcd-kube-api-access-dvn5v" (OuterVolumeSpecName: "kube-api-access-dvn5v") pod "552313d0-57ae-4e50-ae23-d7a568e1afcd" (UID: "552313d0-57ae-4e50-ae23-d7a568e1afcd"). InnerVolumeSpecName "kube-api-access-dvn5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.323152 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552313d0-57ae-4e50-ae23-d7a568e1afcd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "552313d0-57ae-4e50-ae23-d7a568e1afcd" (UID: "552313d0-57ae-4e50-ae23-d7a568e1afcd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.392161 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.414782 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvn5v\" (UniqueName: \"kubernetes.io/projected/552313d0-57ae-4e50-ae23-d7a568e1afcd-kube-api-access-dvn5v\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.414833 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.414846 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552313d0-57ae-4e50-ae23-d7a568e1afcd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.414858 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/552313d0-57ae-4e50-ae23-d7a568e1afcd-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.645925 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f54f59f4f-dkkps"] Dec 08 14:51:25 crc kubenswrapper[4894]: W1208 14:51:25.650180 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc42ba707_9381_4fc9_884d_9cbe5ac31cb9.slice/crio-4fb60a09818a0142633b294c41809d60ed95159f684a842dd9d497da608f8a50 WatchSource:0}: Error finding container 4fb60a09818a0142633b294c41809d60ed95159f684a842dd9d497da608f8a50: Status 404 returned error can't find the container with id 4fb60a09818a0142633b294c41809d60ed95159f684a842dd9d497da608f8a50 Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.892207 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7cr5" event={"ID":"9393c174-85f4-4267-80a1-e7c1075104b5","Type":"ContainerStarted","Data":"425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807"} Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.895116 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bhxcd" event={"ID":"28a7ac74-25ab-4f68-8aee-adbbf5defd99","Type":"ContainerStarted","Data":"3174f184a8addbaf56406751c15c2994c6683b725fc4d94094b3f0adc8cc166d"} Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.897183 4894 generic.go:334] "Generic (PLEG): container finished" podID="552313d0-57ae-4e50-ae23-d7a568e1afcd" containerID="e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884" exitCode=0 Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.897266 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.897723 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" event={"ID":"552313d0-57ae-4e50-ae23-d7a568e1afcd","Type":"ContainerDied","Data":"e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884"} Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.897802 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt" event={"ID":"552313d0-57ae-4e50-ae23-d7a568e1afcd","Type":"ContainerDied","Data":"83a49cb88d8f69508a9f341995c2d19fc6e5ffe1f2cf0be00bcec43b913630f2"} Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.897842 4894 scope.go:117] "RemoveContainer" containerID="e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.899522 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" event={"ID":"c42ba707-9381-4fc9-884d-9cbe5ac31cb9","Type":"ContainerStarted","Data":"9954f49b7f37a84c07d92c32694813d0c139401b8b836b8c800f374323f0bbfc"} Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.899546 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" event={"ID":"c42ba707-9381-4fc9-884d-9cbe5ac31cb9","Type":"ContainerStarted","Data":"4fb60a09818a0142633b294c41809d60ed95159f684a842dd9d497da608f8a50"} Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.900638 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.905356 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.919258 4894 scope.go:117] "RemoveContainer" containerID="e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884" Dec 08 14:51:25 crc kubenswrapper[4894]: E1208 14:51:25.921229 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884\": container with ID starting with e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884 not found: ID does not exist" containerID="e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.921269 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884"} err="failed to get container status \"e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884\": rpc error: code = NotFound desc = could not find container \"e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884\": container with ID starting with e467c6cf955ee6aba8f5441e02cc1bdc0ee5d0e7c229ee7265f4f4e23ab33884 not found: ID does not exist" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.930042 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s7cr5" podStartSLOduration=2.490340016 podStartE2EDuration="4.930014458s" podCreationTimestamp="2025-12-08 14:51:21 +0000 UTC" firstStartedPulling="2025-12-08 14:51:22.856708529 +0000 UTC m=+303.956714644" lastFinishedPulling="2025-12-08 14:51:25.296382971 +0000 UTC m=+306.396389086" observedRunningTime="2025-12-08 14:51:25.912310117 +0000 UTC m=+307.012316242" watchObservedRunningTime="2025-12-08 14:51:25.930014458 +0000 UTC m=+307.030020583" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.933065 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f54f59f4f-dkkps" podStartSLOduration=0.933056999 podStartE2EDuration="933.056999ms" podCreationTimestamp="2025-12-08 14:51:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:51:25.92768014 +0000 UTC m=+307.027686275" watchObservedRunningTime="2025-12-08 14:51:25.933056999 +0000 UTC m=+307.033063134" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.954464 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bhxcd" podStartSLOduration=2.257416402 podStartE2EDuration="4.954441451s" podCreationTimestamp="2025-12-08 14:51:21 +0000 UTC" firstStartedPulling="2025-12-08 14:51:22.844451891 +0000 UTC m=+303.944458006" lastFinishedPulling="2025-12-08 14:51:25.54147694 +0000 UTC m=+306.641483055" observedRunningTime="2025-12-08 14:51:25.951205433 +0000 UTC m=+307.051211568" watchObservedRunningTime="2025-12-08 14:51:25.954441451 +0000 UTC m=+307.054447566" Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.972424 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt"] Dec 08 14:51:25 crc kubenswrapper[4894]: I1208 14:51:25.977315 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-txhpt"] Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.495947 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7"] Dec 08 14:51:26 crc kubenswrapper[4894]: E1208 14:51:26.496426 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552313d0-57ae-4e50-ae23-d7a568e1afcd" containerName="route-controller-manager" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.496439 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="552313d0-57ae-4e50-ae23-d7a568e1afcd" containerName="route-controller-manager" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.496524 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="552313d0-57ae-4e50-ae23-d7a568e1afcd" containerName="route-controller-manager" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.496890 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.500150 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.500411 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.500899 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.500910 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.501237 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.501316 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.513989 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7"] Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.630127 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-config\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.630217 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-client-ca\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.630660 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-serving-cert\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.630739 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9kkp\" (UniqueName: \"kubernetes.io/projected/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-kube-api-access-h9kkp\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.732374 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9kkp\" (UniqueName: \"kubernetes.io/projected/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-kube-api-access-h9kkp\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.732463 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-config\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.732485 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-client-ca\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.732519 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-serving-cert\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.733768 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-client-ca\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.733991 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-config\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.738359 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-serving-cert\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.753483 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9kkp\" (UniqueName: \"kubernetes.io/projected/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-kube-api-access-h9kkp\") pod \"route-controller-manager-5cb4468b7-k9mj7\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:26 crc kubenswrapper[4894]: I1208 14:51:26.816417 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:27 crc kubenswrapper[4894]: I1208 14:51:27.206839 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="552313d0-57ae-4e50-ae23-d7a568e1afcd" path="/var/lib/kubelet/pods/552313d0-57ae-4e50-ae23-d7a568e1afcd/volumes" Dec 08 14:51:27 crc kubenswrapper[4894]: I1208 14:51:27.270749 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7"] Dec 08 14:51:27 crc kubenswrapper[4894]: I1208 14:51:27.924924 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" event={"ID":"65c1cc0d-be3e-4716-b77d-4c4c70c733c2","Type":"ContainerStarted","Data":"6e58a61dcea98fc75541bbc9a206d39a81a4a19593127183e9836f53efb1a617"} Dec 08 14:51:28 crc kubenswrapper[4894]: I1208 14:51:28.930595 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" event={"ID":"65c1cc0d-be3e-4716-b77d-4c4c70c733c2","Type":"ContainerStarted","Data":"e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc"} Dec 08 14:51:28 crc kubenswrapper[4894]: I1208 14:51:28.930954 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:28 crc kubenswrapper[4894]: I1208 14:51:28.935266 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:51:28 crc kubenswrapper[4894]: I1208 14:51:28.943688 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" podStartSLOduration=2.943664808 podStartE2EDuration="2.943664808s" podCreationTimestamp="2025-12-08 14:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:51:28.942642994 +0000 UTC m=+310.042649109" watchObservedRunningTime="2025-12-08 14:51:28.943664808 +0000 UTC m=+310.043670923" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.285458 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.285513 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.332381 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.463269 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.463316 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.502897 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.974751 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qc7vh" Dec 08 14:51:29 crc kubenswrapper[4894]: I1208 14:51:29.977839 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.666577 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.668061 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.704569 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.870114 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.870170 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.916660 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.981083 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bhxcd" Dec 08 14:51:31 crc kubenswrapper[4894]: I1208 14:51:31.989521 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 14:51:55 crc kubenswrapper[4894]: I1208 14:51:55.912693 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m74jd"] Dec 08 14:51:55 crc kubenswrapper[4894]: I1208 14:51:55.913883 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:55 crc kubenswrapper[4894]: I1208 14:51:55.974612 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m74jd"] Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069347 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069399 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgx5c\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-kube-api-access-kgx5c\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069447 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-registry-tls\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069552 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-bound-sa-token\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069588 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-trusted-ca\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069619 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.069641 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-registry-certificates\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.100190 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.171505 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-registry-tls\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.171565 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-bound-sa-token\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.171624 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-trusted-ca\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.171669 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-registry-certificates\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.171714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.171792 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.171852 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgx5c\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-kube-api-access-kgx5c\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.172744 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.173143 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-registry-certificates\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.174018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-trusted-ca\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.177467 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-registry-tls\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.177698 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.186930 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-bound-sa-token\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.197501 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgx5c\" (UniqueName: \"kubernetes.io/projected/5a9c9bc1-e0c0-44e2-b02a-5500f5871acc-kube-api-access-kgx5c\") pod \"image-registry-66df7c8f76-m74jd\" (UID: \"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc\") " pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.229143 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:56 crc kubenswrapper[4894]: I1208 14:51:56.627857 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m74jd"] Dec 08 14:51:57 crc kubenswrapper[4894]: I1208 14:51:57.065239 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" event={"ID":"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc","Type":"ContainerStarted","Data":"6a165a909245cef2577c1675982176fc754717ba681e344a05d9ab6bdb8333c6"} Dec 08 14:51:57 crc kubenswrapper[4894]: I1208 14:51:57.065604 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" event={"ID":"5a9c9bc1-e0c0-44e2-b02a-5500f5871acc","Type":"ContainerStarted","Data":"743e721d6bb7b318ee5b2582e1395170eb23dd84cdef6611ea34e1671d61657c"} Dec 08 14:51:57 crc kubenswrapper[4894]: I1208 14:51:57.065625 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:51:57 crc kubenswrapper[4894]: I1208 14:51:57.082769 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" podStartSLOduration=2.082747054 podStartE2EDuration="2.082747054s" podCreationTimestamp="2025-12-08 14:51:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:51:57.080874653 +0000 UTC m=+338.180880788" watchObservedRunningTime="2025-12-08 14:51:57.082747054 +0000 UTC m=+338.182753169" Dec 08 14:52:07 crc kubenswrapper[4894]: I1208 14:52:07.296924 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:52:07 crc kubenswrapper[4894]: I1208 14:52:07.297415 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:52:16 crc kubenswrapper[4894]: I1208 14:52:16.234131 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-m74jd" Dec 08 14:52:16 crc kubenswrapper[4894]: I1208 14:52:16.282851 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vnpl8"] Dec 08 14:52:37 crc kubenswrapper[4894]: I1208 14:52:37.297310 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:52:37 crc kubenswrapper[4894]: I1208 14:52:37.298066 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:52:41 crc kubenswrapper[4894]: I1208 14:52:41.328021 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" podUID="8a9ef065-72ff-4aa4-a729-c4e9884728e6" containerName="registry" containerID="cri-o://be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225" gracePeriod=30 Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.191679 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.299212 4894 generic.go:334] "Generic (PLEG): container finished" podID="8a9ef065-72ff-4aa4-a729-c4e9884728e6" containerID="be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225" exitCode=0 Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.299254 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" event={"ID":"8a9ef065-72ff-4aa4-a729-c4e9884728e6","Type":"ContainerDied","Data":"be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225"} Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.299280 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.299305 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vnpl8" event={"ID":"8a9ef065-72ff-4aa4-a729-c4e9884728e6","Type":"ContainerDied","Data":"66db9056d3476b9cb2ceec84a72a0353eb08f2859b16e39578cc7d26f47f707d"} Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.299324 4894 scope.go:117] "RemoveContainer" containerID="be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.313724 4894 scope.go:117] "RemoveContainer" containerID="be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225" Dec 08 14:52:42 crc kubenswrapper[4894]: E1208 14:52:42.314174 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225\": container with ID starting with be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225 not found: ID does not exist" containerID="be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.314234 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225"} err="failed to get container status \"be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225\": rpc error: code = NotFound desc = could not find container \"be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225\": container with ID starting with be8c6fed670fea940b143342ef82826c6927889166b07c25e801b3cd342ee225 not found: ID does not exist" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.326458 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a9ef065-72ff-4aa4-a729-c4e9884728e6-ca-trust-extracted\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.326609 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-trusted-ca\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.326643 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a9ef065-72ff-4aa4-a729-c4e9884728e6-installation-pull-secrets\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.326780 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.326879 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-bound-sa-token\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.326940 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-tls\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.327050 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-certificates\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.327113 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk8qv\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-kube-api-access-sk8qv\") pod \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\" (UID: \"8a9ef065-72ff-4aa4-a729-c4e9884728e6\") " Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.327751 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.328288 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.333729 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.333845 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-kube-api-access-sk8qv" (OuterVolumeSpecName: "kube-api-access-sk8qv") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "kube-api-access-sk8qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.333867 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.334378 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a9ef065-72ff-4aa4-a729-c4e9884728e6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.335102 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.348633 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a9ef065-72ff-4aa4-a729-c4e9884728e6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8a9ef065-72ff-4aa4-a729-c4e9884728e6" (UID: "8a9ef065-72ff-4aa4-a729-c4e9884728e6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.428054 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.428082 4894 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8a9ef065-72ff-4aa4-a729-c4e9884728e6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.428092 4894 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.428100 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.428109 4894 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8a9ef065-72ff-4aa4-a729-c4e9884728e6-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.428117 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk8qv\" (UniqueName: \"kubernetes.io/projected/8a9ef065-72ff-4aa4-a729-c4e9884728e6-kube-api-access-sk8qv\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.428125 4894 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8a9ef065-72ff-4aa4-a729-c4e9884728e6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.627572 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vnpl8"] Dec 08 14:52:42 crc kubenswrapper[4894]: I1208 14:52:42.633504 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vnpl8"] Dec 08 14:52:43 crc kubenswrapper[4894]: I1208 14:52:43.203933 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a9ef065-72ff-4aa4-a729-c4e9884728e6" path="/var/lib/kubelet/pods/8a9ef065-72ff-4aa4-a729-c4e9884728e6/volumes" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.393528 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7"] Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.393727 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" podUID="65c1cc0d-be3e-4716-b77d-4c4c70c733c2" containerName="route-controller-manager" containerID="cri-o://e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc" gracePeriod=30 Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.733613 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.855423 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9kkp\" (UniqueName: \"kubernetes.io/projected/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-kube-api-access-h9kkp\") pod \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.855557 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-client-ca\") pod \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.855594 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-config\") pod \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.855643 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-serving-cert\") pod \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\" (UID: \"65c1cc0d-be3e-4716-b77d-4c4c70c733c2\") " Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.856430 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-config" (OuterVolumeSpecName: "config") pod "65c1cc0d-be3e-4716-b77d-4c4c70c733c2" (UID: "65c1cc0d-be3e-4716-b77d-4c4c70c733c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.856421 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-client-ca" (OuterVolumeSpecName: "client-ca") pod "65c1cc0d-be3e-4716-b77d-4c4c70c733c2" (UID: "65c1cc0d-be3e-4716-b77d-4c4c70c733c2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.858934 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "65c1cc0d-be3e-4716-b77d-4c4c70c733c2" (UID: "65c1cc0d-be3e-4716-b77d-4c4c70c733c2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.859421 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-kube-api-access-h9kkp" (OuterVolumeSpecName: "kube-api-access-h9kkp") pod "65c1cc0d-be3e-4716-b77d-4c4c70c733c2" (UID: "65c1cc0d-be3e-4716-b77d-4c4c70c733c2"). InnerVolumeSpecName "kube-api-access-h9kkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.957611 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9kkp\" (UniqueName: \"kubernetes.io/projected/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-kube-api-access-h9kkp\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.957656 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.957668 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:44 crc kubenswrapper[4894]: I1208 14:52:44.957678 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65c1cc0d-be3e-4716-b77d-4c4c70c733c2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.318525 4894 generic.go:334] "Generic (PLEG): container finished" podID="65c1cc0d-be3e-4716-b77d-4c4c70c733c2" containerID="e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc" exitCode=0 Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.318591 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" event={"ID":"65c1cc0d-be3e-4716-b77d-4c4c70c733c2","Type":"ContainerDied","Data":"e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc"} Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.318650 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.318683 4894 scope.go:117] "RemoveContainer" containerID="e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.318665 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7" event={"ID":"65c1cc0d-be3e-4716-b77d-4c4c70c733c2","Type":"ContainerDied","Data":"6e58a61dcea98fc75541bbc9a206d39a81a4a19593127183e9836f53efb1a617"} Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.341545 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7"] Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.342206 4894 scope.go:117] "RemoveContainer" containerID="e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc" Dec 08 14:52:45 crc kubenswrapper[4894]: E1208 14:52:45.342776 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc\": container with ID starting with e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc not found: ID does not exist" containerID="e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.342834 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc"} err="failed to get container status \"e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc\": rpc error: code = NotFound desc = could not find container \"e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc\": container with ID starting with e02c22adfffbb9d3464c33a459b78e70876993ad1f8c35da4caa543bdd323bcc not found: ID does not exist" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.344643 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cb4468b7-k9mj7"] Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.543643 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh"] Dec 08 14:52:45 crc kubenswrapper[4894]: E1208 14:52:45.544432 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65c1cc0d-be3e-4716-b77d-4c4c70c733c2" containerName="route-controller-manager" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.544446 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="65c1cc0d-be3e-4716-b77d-4c4c70c733c2" containerName="route-controller-manager" Dec 08 14:52:45 crc kubenswrapper[4894]: E1208 14:52:45.544457 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a9ef065-72ff-4aa4-a729-c4e9884728e6" containerName="registry" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.544463 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a9ef065-72ff-4aa4-a729-c4e9884728e6" containerName="registry" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.544554 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="65c1cc0d-be3e-4716-b77d-4c4c70c733c2" containerName="route-controller-manager" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.544574 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a9ef065-72ff-4aa4-a729-c4e9884728e6" containerName="registry" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.544945 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.547415 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.547633 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.547939 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.548176 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.548316 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.548720 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.560176 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh"] Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.580724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-client-ca\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.580938 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-config\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.681873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-config\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.681929 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-serving-cert\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.681974 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-client-ca\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.682022 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cffkf\" (UniqueName: \"kubernetes.io/projected/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-kube-api-access-cffkf\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.683042 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-config\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.683400 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-client-ca\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.783317 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cffkf\" (UniqueName: \"kubernetes.io/projected/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-kube-api-access-cffkf\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.783409 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-serving-cert\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.788694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-serving-cert\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.798660 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cffkf\" (UniqueName: \"kubernetes.io/projected/7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e-kube-api-access-cffkf\") pod \"route-controller-manager-66689d8ff-vkzmh\" (UID: \"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e\") " pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:45 crc kubenswrapper[4894]: I1208 14:52:45.867956 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:46 crc kubenswrapper[4894]: I1208 14:52:46.057901 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh"] Dec 08 14:52:46 crc kubenswrapper[4894]: I1208 14:52:46.327025 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" event={"ID":"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e","Type":"ContainerStarted","Data":"0fc1c95b0075d38d8816a72bc1820362585eb69226282ba8b37608c18c5a593d"} Dec 08 14:52:46 crc kubenswrapper[4894]: I1208 14:52:46.327073 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" event={"ID":"7b78cdf5-492b-4bbd-ae0e-a0cca6514d2e","Type":"ContainerStarted","Data":"64eec2b56d0cf62a5ce91095db0b66cd4bd477a6a01a75ed016ad245a032831c"} Dec 08 14:52:46 crc kubenswrapper[4894]: I1208 14:52:46.327326 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:46 crc kubenswrapper[4894]: I1208 14:52:46.352736 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" podStartSLOduration=2.352713929 podStartE2EDuration="2.352713929s" podCreationTimestamp="2025-12-08 14:52:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:52:46.344658034 +0000 UTC m=+387.444664149" watchObservedRunningTime="2025-12-08 14:52:46.352713929 +0000 UTC m=+387.452720044" Dec 08 14:52:46 crc kubenswrapper[4894]: I1208 14:52:46.712607 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66689d8ff-vkzmh" Dec 08 14:52:47 crc kubenswrapper[4894]: I1208 14:52:47.203851 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65c1cc0d-be3e-4716-b77d-4c4c70c733c2" path="/var/lib/kubelet/pods/65c1cc0d-be3e-4716-b77d-4c4c70c733c2/volumes" Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.297262 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.297767 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.297857 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.298423 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"446f34182251fad2ef136856c085177483a1e0666f7e36e3916d271f2f7efad9"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.298478 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://446f34182251fad2ef136856c085177483a1e0666f7e36e3916d271f2f7efad9" gracePeriod=600 Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.430571 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="446f34182251fad2ef136856c085177483a1e0666f7e36e3916d271f2f7efad9" exitCode=0 Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.430625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"446f34182251fad2ef136856c085177483a1e0666f7e36e3916d271f2f7efad9"} Dec 08 14:53:07 crc kubenswrapper[4894]: I1208 14:53:07.430665 4894 scope.go:117] "RemoveContainer" containerID="6a3e26ec97b2ec49ae32982b1aa7fd1e903f1370d336429d7ac92ddf514ce1a6" Dec 08 14:53:08 crc kubenswrapper[4894]: I1208 14:53:08.438167 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"bcf05c5956a90ca1ef0b1249396e38947db07b49d45785534a1f514e4747c039"} Dec 08 14:55:07 crc kubenswrapper[4894]: I1208 14:55:07.296961 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:55:07 crc kubenswrapper[4894]: I1208 14:55:07.297489 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:55:37 crc kubenswrapper[4894]: I1208 14:55:37.296534 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:55:37 crc kubenswrapper[4894]: I1208 14:55:37.297958 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:56:07 crc kubenswrapper[4894]: I1208 14:56:07.296804 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:56:07 crc kubenswrapper[4894]: I1208 14:56:07.297323 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:56:07 crc kubenswrapper[4894]: I1208 14:56:07.297366 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:56:07 crc kubenswrapper[4894]: I1208 14:56:07.297839 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bcf05c5956a90ca1ef0b1249396e38947db07b49d45785534a1f514e4747c039"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 14:56:07 crc kubenswrapper[4894]: I1208 14:56:07.297894 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://bcf05c5956a90ca1ef0b1249396e38947db07b49d45785534a1f514e4747c039" gracePeriod=600 Dec 08 14:56:08 crc kubenswrapper[4894]: I1208 14:56:08.385440 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="bcf05c5956a90ca1ef0b1249396e38947db07b49d45785534a1f514e4747c039" exitCode=0 Dec 08 14:56:08 crc kubenswrapper[4894]: I1208 14:56:08.385523 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"bcf05c5956a90ca1ef0b1249396e38947db07b49d45785534a1f514e4747c039"} Dec 08 14:56:08 crc kubenswrapper[4894]: I1208 14:56:08.385739 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"dfc08546d7b8c2b8b1aa3ec04284ba529dbbd4025e73389fcf8cb4e001930019"} Dec 08 14:56:08 crc kubenswrapper[4894]: I1208 14:56:08.385759 4894 scope.go:117] "RemoveContainer" containerID="446f34182251fad2ef136856c085177483a1e0666f7e36e3916d271f2f7efad9" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.117067 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qfw89"] Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.118185 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.122149 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tbs5t" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.122178 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.122314 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.123295 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nzrw5"] Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.124107 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nzrw5" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.126186 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-xvktq" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.130273 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qfw89"] Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.145238 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hzvvg"] Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.149270 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.151278 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hzvvg"] Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.153022 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mb655" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.154643 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nzrw5"] Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.313882 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnk5v\" (UniqueName: \"kubernetes.io/projected/960dc174-5cd7-4931-94f7-79871062fcb9-kube-api-access-gnk5v\") pod \"cert-manager-cainjector-7f985d654d-qfw89\" (UID: \"960dc174-5cd7-4931-94f7-79871062fcb9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.313989 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdrvl\" (UniqueName: \"kubernetes.io/projected/0aa35d3c-6f53-415d-8a11-f2e8cc6298fb-kube-api-access-mdrvl\") pod \"cert-manager-5b446d88c5-nzrw5\" (UID: \"0aa35d3c-6f53-415d-8a11-f2e8cc6298fb\") " pod="cert-manager/cert-manager-5b446d88c5-nzrw5" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.314046 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jgxs\" (UniqueName: \"kubernetes.io/projected/7774fcfd-7cfb-4b62-8e29-48c16c5d0337-kube-api-access-8jgxs\") pod \"cert-manager-webhook-5655c58dd6-hzvvg\" (UID: \"7774fcfd-7cfb-4b62-8e29-48c16c5d0337\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.415214 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnk5v\" (UniqueName: \"kubernetes.io/projected/960dc174-5cd7-4931-94f7-79871062fcb9-kube-api-access-gnk5v\") pod \"cert-manager-cainjector-7f985d654d-qfw89\" (UID: \"960dc174-5cd7-4931-94f7-79871062fcb9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.415637 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdrvl\" (UniqueName: \"kubernetes.io/projected/0aa35d3c-6f53-415d-8a11-f2e8cc6298fb-kube-api-access-mdrvl\") pod \"cert-manager-5b446d88c5-nzrw5\" (UID: \"0aa35d3c-6f53-415d-8a11-f2e8cc6298fb\") " pod="cert-manager/cert-manager-5b446d88c5-nzrw5" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.415926 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jgxs\" (UniqueName: \"kubernetes.io/projected/7774fcfd-7cfb-4b62-8e29-48c16c5d0337-kube-api-access-8jgxs\") pod \"cert-manager-webhook-5655c58dd6-hzvvg\" (UID: \"7774fcfd-7cfb-4b62-8e29-48c16c5d0337\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.432513 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdrvl\" (UniqueName: \"kubernetes.io/projected/0aa35d3c-6f53-415d-8a11-f2e8cc6298fb-kube-api-access-mdrvl\") pod \"cert-manager-5b446d88c5-nzrw5\" (UID: \"0aa35d3c-6f53-415d-8a11-f2e8cc6298fb\") " pod="cert-manager/cert-manager-5b446d88c5-nzrw5" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.432913 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jgxs\" (UniqueName: \"kubernetes.io/projected/7774fcfd-7cfb-4b62-8e29-48c16c5d0337-kube-api-access-8jgxs\") pod \"cert-manager-webhook-5655c58dd6-hzvvg\" (UID: \"7774fcfd-7cfb-4b62-8e29-48c16c5d0337\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.433310 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnk5v\" (UniqueName: \"kubernetes.io/projected/960dc174-5cd7-4931-94f7-79871062fcb9-kube-api-access-gnk5v\") pod \"cert-manager-cainjector-7f985d654d-qfw89\" (UID: \"960dc174-5cd7-4931-94f7-79871062fcb9\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.436608 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.443644 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nzrw5" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.467768 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.646662 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nzrw5"] Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.656321 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.699840 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qfw89"] Dec 08 14:56:46 crc kubenswrapper[4894]: W1208 14:56:46.702683 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod960dc174_5cd7_4931_94f7_79871062fcb9.slice/crio-ce08967f2b858c8582b53dd04613d0bc158cbe194ab9aac2aa2be5d199b8e5ca WatchSource:0}: Error finding container ce08967f2b858c8582b53dd04613d0bc158cbe194ab9aac2aa2be5d199b8e5ca: Status 404 returned error can't find the container with id ce08967f2b858c8582b53dd04613d0bc158cbe194ab9aac2aa2be5d199b8e5ca Dec 08 14:56:46 crc kubenswrapper[4894]: I1208 14:56:46.726089 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hzvvg"] Dec 08 14:56:46 crc kubenswrapper[4894]: W1208 14:56:46.732948 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7774fcfd_7cfb_4b62_8e29_48c16c5d0337.slice/crio-94ec3c69feedc1a3a74d8e8f38a02e6bcefb869f678be69f2ae38cf2d401c85e WatchSource:0}: Error finding container 94ec3c69feedc1a3a74d8e8f38a02e6bcefb869f678be69f2ae38cf2d401c85e: Status 404 returned error can't find the container with id 94ec3c69feedc1a3a74d8e8f38a02e6bcefb869f678be69f2ae38cf2d401c85e Dec 08 14:56:47 crc kubenswrapper[4894]: I1208 14:56:47.611250 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nzrw5" event={"ID":"0aa35d3c-6f53-415d-8a11-f2e8cc6298fb","Type":"ContainerStarted","Data":"36ac102a62738a767f754674bdf2a2a34bbfb9e4830a8ea2d14b3b39a8459ebb"} Dec 08 14:56:47 crc kubenswrapper[4894]: I1208 14:56:47.612074 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" event={"ID":"7774fcfd-7cfb-4b62-8e29-48c16c5d0337","Type":"ContainerStarted","Data":"94ec3c69feedc1a3a74d8e8f38a02e6bcefb869f678be69f2ae38cf2d401c85e"} Dec 08 14:56:47 crc kubenswrapper[4894]: I1208 14:56:47.612779 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" event={"ID":"960dc174-5cd7-4931-94f7-79871062fcb9","Type":"ContainerStarted","Data":"ce08967f2b858c8582b53dd04613d0bc158cbe194ab9aac2aa2be5d199b8e5ca"} Dec 08 14:56:50 crc kubenswrapper[4894]: I1208 14:56:50.631275 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nzrw5" event={"ID":"0aa35d3c-6f53-415d-8a11-f2e8cc6298fb","Type":"ContainerStarted","Data":"dfdbc6630becb8b56ad1f15d821923807bf22ef561177bd13ca910d33286aba5"} Dec 08 14:56:50 crc kubenswrapper[4894]: I1208 14:56:50.634335 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" event={"ID":"7774fcfd-7cfb-4b62-8e29-48c16c5d0337","Type":"ContainerStarted","Data":"4b7afead35850373c43727bff337b930858602477d31a99d0a98b939b877b7e8"} Dec 08 14:56:50 crc kubenswrapper[4894]: I1208 14:56:50.634763 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" Dec 08 14:56:50 crc kubenswrapper[4894]: I1208 14:56:50.636769 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" event={"ID":"960dc174-5cd7-4931-94f7-79871062fcb9","Type":"ContainerStarted","Data":"573289bccc7618be684c0eec07c63db57e1a1a7c61d590220b1690a517896792"} Dec 08 14:56:50 crc kubenswrapper[4894]: I1208 14:56:50.660150 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-nzrw5" podStartSLOduration=1.7656052089999998 podStartE2EDuration="4.660134646s" podCreationTimestamp="2025-12-08 14:56:46 +0000 UTC" firstStartedPulling="2025-12-08 14:56:46.656105617 +0000 UTC m=+627.756111732" lastFinishedPulling="2025-12-08 14:56:49.550635054 +0000 UTC m=+630.650641169" observedRunningTime="2025-12-08 14:56:50.65744662 +0000 UTC m=+631.757452765" watchObservedRunningTime="2025-12-08 14:56:50.660134646 +0000 UTC m=+631.760140761" Dec 08 14:56:50 crc kubenswrapper[4894]: I1208 14:56:50.689582 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qfw89" podStartSLOduration=1.907342589 podStartE2EDuration="4.689558447s" podCreationTimestamp="2025-12-08 14:56:46 +0000 UTC" firstStartedPulling="2025-12-08 14:56:46.705639371 +0000 UTC m=+627.805645486" lastFinishedPulling="2025-12-08 14:56:49.487855229 +0000 UTC m=+630.587861344" observedRunningTime="2025-12-08 14:56:50.685108364 +0000 UTC m=+631.785114479" watchObservedRunningTime="2025-12-08 14:56:50.689558447 +0000 UTC m=+631.789564572" Dec 08 14:56:50 crc kubenswrapper[4894]: I1208 14:56:50.731901 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" podStartSLOduration=1.978821911 podStartE2EDuration="4.731877018s" podCreationTimestamp="2025-12-08 14:56:46 +0000 UTC" firstStartedPulling="2025-12-08 14:56:46.73442045 +0000 UTC m=+627.834426565" lastFinishedPulling="2025-12-08 14:56:49.487475557 +0000 UTC m=+630.587481672" observedRunningTime="2025-12-08 14:56:50.727732996 +0000 UTC m=+631.827739111" watchObservedRunningTime="2025-12-08 14:56:50.731877018 +0000 UTC m=+631.831883143" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.430777 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c25bk"] Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.435533 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-controller" containerID="cri-o://39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.435704 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.435722 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="northd" containerID="cri-o://fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.435753 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-node" containerID="cri-o://cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.435800 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-acl-logging" containerID="cri-o://118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.436084 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="sbdb" containerID="cri-o://f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.436139 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="nbdb" containerID="cri-o://307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.468354 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" containerID="cri-o://533baecbcd1f2894c4dd1141d33b4fb719b2d10aff57238aa5684d387568d9f6" gracePeriod=30 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.470656 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-hzvvg" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.669325 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovnkube-controller/3.log" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.671268 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovn-acl-logging/0.log" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.671691 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovn-controller/0.log" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672061 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="533baecbcd1f2894c4dd1141d33b4fb719b2d10aff57238aa5684d387568d9f6" exitCode=0 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672086 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b" exitCode=0 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672093 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8" exitCode=0 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672101 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a" exitCode=0 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672109 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d" exitCode=0 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672117 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f" exitCode=0 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672123 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354" exitCode=143 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672131 4894 generic.go:334] "Generic (PLEG): container finished" podID="fb212a28-36c5-440e-8965-986352c5d3ea" containerID="39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787" exitCode=143 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672131 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"533baecbcd1f2894c4dd1141d33b4fb719b2d10aff57238aa5684d387568d9f6"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672192 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672206 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672218 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672239 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.672195 4894 scope.go:117] "RemoveContainer" containerID="d581ada418c1b8ec89a0682d99b96dfe58b9b97a8e9beafd186048638114a189" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.674061 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/2.log" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.674659 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/1.log" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.674757 4894 generic.go:334] "Generic (PLEG): container finished" podID="be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235" containerID="49c32397a748d68b9ff9dc6f0fddb18ac746953d7dce6f7c93da63f6b3337ba7" exitCode=2 Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.674863 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerDied","Data":"49c32397a748d68b9ff9dc6f0fddb18ac746953d7dce6f7c93da63f6b3337ba7"} Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.675398 4894 scope.go:117] "RemoveContainer" containerID="49c32397a748d68b9ff9dc6f0fddb18ac746953d7dce6f7c93da63f6b3337ba7" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.675648 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jtx8z_openshift-multus(be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235)\"" pod="openshift-multus/multus-jtx8z" podUID="be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.721247 4894 scope.go:117] "RemoveContainer" containerID="8110a8a29b02ac9a8ba7ba4875f7726388f8c944e6a55afe74f7e801b735e937" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.746033 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovn-acl-logging/0.log" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.746491 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovn-controller/0.log" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.747071 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.800797 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9lsz2"] Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801069 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801088 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801103 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="nbdb" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801112 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="nbdb" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801122 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801130 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801140 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-acl-logging" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801147 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-acl-logging" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801163 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-ovn-metrics" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801174 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-ovn-metrics" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801187 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-node" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801194 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-node" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801202 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="northd" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801210 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="northd" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801221 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kubecfg-setup" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801228 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kubecfg-setup" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801241 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801248 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801257 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="sbdb" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801264 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="sbdb" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801271 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801279 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801389 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801399 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="northd" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801410 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-ovn-metrics" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801417 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="sbdb" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801426 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-acl-logging" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801436 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801444 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801455 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="kube-rbac-proxy-node" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801464 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovn-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801475 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801508 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="nbdb" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801626 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801637 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: E1208 14:56:56.801648 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801655 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.801766 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" containerName="ovnkube-controller" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.803394 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862250 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-ovn-kubernetes\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862295 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-etc-openvswitch\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862319 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-kubelet\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862357 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862357 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862426 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862480 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-systemd-units\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862511 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862694 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-netns\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862761 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-log-socket\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862770 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862798 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-env-overrides\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862831 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-slash\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-log-socket" (OuterVolumeSpecName: "log-socket") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862863 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-netd\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862903 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-slash" (OuterVolumeSpecName: "host-slash") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862907 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-bin\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862922 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862925 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862939 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb212a28-36c5-440e-8965-986352c5d3ea-ovn-node-metrics-cert\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862966 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-script-lib\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.862986 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-var-lib-openvswitch\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6qpk\" (UniqueName: \"kubernetes.io/projected/fb212a28-36c5-440e-8965-986352c5d3ea-kube-api-access-k6qpk\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863029 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-config\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863050 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863054 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-systemd\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863077 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-node-log\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863109 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863133 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-openvswitch\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863153 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-ovn\") pod \"fb212a28-36c5-440e-8965-986352c5d3ea\" (UID: \"fb212a28-36c5-440e-8965-986352c5d3ea\") " Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863232 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863259 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-slash\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863282 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-var-lib-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863302 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-run-ovn-kubernetes\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863324 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863331 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-run-netns\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-log-socket\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863458 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-systemd\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863482 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-cni-bin\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863487 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863503 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863517 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863531 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-ovnkube-config\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863542 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863544 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863557 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/532baa53-9958-4e83-89b0-e81b43e1133d-ovn-node-metrics-cert\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863575 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-node-log" (OuterVolumeSpecName: "node-log") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863735 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln5xm\" (UniqueName: \"kubernetes.io/projected/532baa53-9958-4e83-89b0-e81b43e1133d-kube-api-access-ln5xm\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863765 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-kubelet\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863806 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-systemd-units\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863843 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-env-overrides\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863861 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863890 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-ovnkube-script-lib\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863907 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-ovn\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863929 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-etc-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863965 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-node-log\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.863981 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-cni-netd\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864045 4894 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-log-socket\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864059 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864071 4894 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-slash\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864082 4894 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864094 4894 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864107 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864119 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb212a28-36c5-440e-8965-986352c5d3ea-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864129 4894 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864140 4894 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-node-log\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864152 4894 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864163 4894 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864175 4894 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864186 4894 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864197 4894 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864207 4894 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864218 4894 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.864229 4894 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.867736 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb212a28-36c5-440e-8965-986352c5d3ea-kube-api-access-k6qpk" (OuterVolumeSpecName: "kube-api-access-k6qpk") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "kube-api-access-k6qpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.868062 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb212a28-36c5-440e-8965-986352c5d3ea-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.874978 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "fb212a28-36c5-440e-8965-986352c5d3ea" (UID: "fb212a28-36c5-440e-8965-986352c5d3ea"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.964646 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-log-socket\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.964743 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-log-socket\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966020 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-systemd\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966099 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-systemd\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966122 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-cni-bin\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966197 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-ovnkube-config\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966241 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/532baa53-9958-4e83-89b0-e81b43e1133d-ovn-node-metrics-cert\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966270 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-cni-bin\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966370 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln5xm\" (UniqueName: \"kubernetes.io/projected/532baa53-9958-4e83-89b0-e81b43e1133d-kube-api-access-ln5xm\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-kubelet\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966437 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-systemd-units\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966456 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-env-overrides\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966472 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966493 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-ovnkube-script-lib\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966509 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-ovn\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966508 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-kubelet\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966530 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-etc-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-etc-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966574 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-ovn\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966581 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-systemd-units\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966641 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-node-log\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.966625 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-node-log\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967118 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-ovnkube-config\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967127 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-cni-netd\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967156 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-cni-netd\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967153 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-env-overrides\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967228 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967263 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-slash\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967291 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-var-lib-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967313 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-run-ovn-kubernetes\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967350 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-run-netns\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967385 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-var-lib-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967397 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-run-ovn-kubernetes\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-slash\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967421 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-host-run-netns\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/532baa53-9958-4e83-89b0-e81b43e1133d-run-openvswitch\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967461 4894 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fb212a28-36c5-440e-8965-986352c5d3ea-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967480 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb212a28-36c5-440e-8965-986352c5d3ea-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967551 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6qpk\" (UniqueName: \"kubernetes.io/projected/fb212a28-36c5-440e-8965-986352c5d3ea-kube-api-access-k6qpk\") on node \"crc\" DevicePath \"\"" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.967561 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/532baa53-9958-4e83-89b0-e81b43e1133d-ovnkube-script-lib\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.971535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/532baa53-9958-4e83-89b0-e81b43e1133d-ovn-node-metrics-cert\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:56 crc kubenswrapper[4894]: I1208 14:56:56.982494 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln5xm\" (UniqueName: \"kubernetes.io/projected/532baa53-9958-4e83-89b0-e81b43e1133d-kube-api-access-ln5xm\") pod \"ovnkube-node-9lsz2\" (UID: \"532baa53-9958-4e83-89b0-e81b43e1133d\") " pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.115580 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.682356 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovn-acl-logging/0.log" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.683465 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c25bk_fb212a28-36c5-440e-8965-986352c5d3ea/ovn-controller/0.log" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.683956 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" event={"ID":"fb212a28-36c5-440e-8965-986352c5d3ea","Type":"ContainerDied","Data":"06a57e2fb05e0b17ee1ffb63945245145b5f09f6faf6b17b6446ba18f3c515b0"} Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.684001 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c25bk" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.684005 4894 scope.go:117] "RemoveContainer" containerID="533baecbcd1f2894c4dd1141d33b4fb719b2d10aff57238aa5684d387568d9f6" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.686311 4894 generic.go:334] "Generic (PLEG): container finished" podID="532baa53-9958-4e83-89b0-e81b43e1133d" containerID="80dcb0c7ee5358f9e387d5371a938aaee2a589e739f1c9f92c460ad8a6acc369" exitCode=0 Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.686360 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerDied","Data":"80dcb0c7ee5358f9e387d5371a938aaee2a589e739f1c9f92c460ad8a6acc369"} Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.686391 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"8420ddadc789488067c91c38e963b1038c3625676a61d0b490817767622e6828"} Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.689427 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/2.log" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.712282 4894 scope.go:117] "RemoveContainer" containerID="f48b16ecb1e7b675a8e5c84316c384473785f5e9c1eb8fae3acb7814fc53806b" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.728832 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c25bk"] Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.734463 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c25bk"] Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.745519 4894 scope.go:117] "RemoveContainer" containerID="307de7370902069e2f32f529a3fbebba699fce34abe70dd3a22c5c70d1d930b8" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.764133 4894 scope.go:117] "RemoveContainer" containerID="fc111ba60e622f0e712d714f74c755f3b1885ebd9b695e2f73f8973a6a07e06a" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.782483 4894 scope.go:117] "RemoveContainer" containerID="4736f42ae313cb369f1be194829dff69e8a756f356bb02a53468391a49b9d78d" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.800666 4894 scope.go:117] "RemoveContainer" containerID="cca50e8eacbf25da8213829f5d0668cf8c579f0ff6dadf2c6369af240b5de03f" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.822606 4894 scope.go:117] "RemoveContainer" containerID="118f5f7d0cf5c44427df41d5440106bb001e1b5b8fba02740344372d706fc354" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.840077 4894 scope.go:117] "RemoveContainer" containerID="39854c5d3011f015e395ff87cd74baa137646cf3bd479587462b7bba2f59e787" Dec 08 14:56:57 crc kubenswrapper[4894]: I1208 14:56:57.875786 4894 scope.go:117] "RemoveContainer" containerID="6b2a02afacfd135aaed586b236eed1e3a75addddf77e95701824b9d6f9c75f95" Dec 08 14:56:58 crc kubenswrapper[4894]: I1208 14:56:58.697758 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"b25b2381594e46f63ae05c61cf1db90f1df9b29c7391120e5db46c5737697717"} Dec 08 14:56:58 crc kubenswrapper[4894]: I1208 14:56:58.698782 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"51d6c512bb21f3dc11017d77d28de1d4cc59312f716a720d7f2aa026efa8d0ca"} Dec 08 14:56:58 crc kubenswrapper[4894]: I1208 14:56:58.698897 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"04ab77c9c14ea6b2c9caa9647e2f8e1dcbdb43f580d46a8058a10f53b9301833"} Dec 08 14:56:58 crc kubenswrapper[4894]: I1208 14:56:58.698964 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"a3432c5127703d14dd22eb205b6ed942f221852219955046734277543bf6395a"} Dec 08 14:56:58 crc kubenswrapper[4894]: I1208 14:56:58.699021 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"1d5274f93730c6dd9944d04fac04dd5b88d6f889a5c45c1aa044e7e9db8f8623"} Dec 08 14:56:58 crc kubenswrapper[4894]: I1208 14:56:58.699081 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"23bdbbeeb8653232fd4460047c7d5acbe86bac62fae08f5ed3a90e17992a1471"} Dec 08 14:56:59 crc kubenswrapper[4894]: I1208 14:56:59.206733 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb212a28-36c5-440e-8965-986352c5d3ea" path="/var/lib/kubelet/pods/fb212a28-36c5-440e-8965-986352c5d3ea/volumes" Dec 08 14:57:00 crc kubenswrapper[4894]: I1208 14:57:00.710217 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"4a1416ba516da33e2111071fdb01dabe3cb079d3eead7328c8b9eb9924ccd15f"} Dec 08 14:57:03 crc kubenswrapper[4894]: I1208 14:57:03.730788 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" event={"ID":"532baa53-9958-4e83-89b0-e81b43e1133d","Type":"ContainerStarted","Data":"7ccfe812d664daf2f2e1bafba15f17b1646aa4fb49b7da3567b27403a4674233"} Dec 08 14:57:03 crc kubenswrapper[4894]: I1208 14:57:03.731150 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:57:03 crc kubenswrapper[4894]: I1208 14:57:03.731186 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:57:03 crc kubenswrapper[4894]: I1208 14:57:03.731304 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:57:03 crc kubenswrapper[4894]: I1208 14:57:03.764140 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:57:03 crc kubenswrapper[4894]: I1208 14:57:03.764225 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:57:03 crc kubenswrapper[4894]: I1208 14:57:03.766656 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" podStartSLOduration=7.766641519 podStartE2EDuration="7.766641519s" podCreationTimestamp="2025-12-08 14:56:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:57:03.764495361 +0000 UTC m=+644.864501476" watchObservedRunningTime="2025-12-08 14:57:03.766641519 +0000 UTC m=+644.866647634" Dec 08 14:57:09 crc kubenswrapper[4894]: I1208 14:57:09.199671 4894 scope.go:117] "RemoveContainer" containerID="49c32397a748d68b9ff9dc6f0fddb18ac746953d7dce6f7c93da63f6b3337ba7" Dec 08 14:57:09 crc kubenswrapper[4894]: E1208 14:57:09.200518 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jtx8z_openshift-multus(be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235)\"" pod="openshift-multus/multus-jtx8z" podUID="be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235" Dec 08 14:57:23 crc kubenswrapper[4894]: I1208 14:57:23.197773 4894 scope.go:117] "RemoveContainer" containerID="49c32397a748d68b9ff9dc6f0fddb18ac746953d7dce6f7c93da63f6b3337ba7" Dec 08 14:57:23 crc kubenswrapper[4894]: I1208 14:57:23.829990 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jtx8z_be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235/kube-multus/2.log" Dec 08 14:57:23 crc kubenswrapper[4894]: I1208 14:57:23.830361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jtx8z" event={"ID":"be0cbbcc-9c4b-4ba8-87d2-f5dc13e57235","Type":"ContainerStarted","Data":"0c072dc2f1920609ee98f4b3a4307d5f363908a19212edbf19f2627148cc647f"} Dec 08 14:57:27 crc kubenswrapper[4894]: I1208 14:57:27.145394 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9lsz2" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.744233 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6"] Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.745792 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.752538 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.753273 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6"] Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.836369 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.836433 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79djf\" (UniqueName: \"kubernetes.io/projected/383ac257-ce50-4d47-8298-799521c308f6-kube-api-access-79djf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.836507 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.937428 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79djf\" (UniqueName: \"kubernetes.io/projected/383ac257-ce50-4d47-8298-799521c308f6-kube-api-access-79djf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.937741 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.937939 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.938257 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.938542 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:35 crc kubenswrapper[4894]: I1208 14:57:35.961616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79djf\" (UniqueName: \"kubernetes.io/projected/383ac257-ce50-4d47-8298-799521c308f6-kube-api-access-79djf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:36 crc kubenswrapper[4894]: I1208 14:57:36.060771 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:36 crc kubenswrapper[4894]: I1208 14:57:36.244499 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6"] Dec 08 14:57:36 crc kubenswrapper[4894]: I1208 14:57:36.898703 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" event={"ID":"383ac257-ce50-4d47-8298-799521c308f6","Type":"ContainerStarted","Data":"d6564e0df45946296e04c44a8a8da5d1dced422fcc301ebcfe7206d704884a1f"} Dec 08 14:57:37 crc kubenswrapper[4894]: I1208 14:57:37.907903 4894 generic.go:334] "Generic (PLEG): container finished" podID="383ac257-ce50-4d47-8298-799521c308f6" containerID="bbabd7db73c749b6bfbaf48e3f39c48b30f6a6fb6dbedff78c3bf3edadaa1874" exitCode=0 Dec 08 14:57:37 crc kubenswrapper[4894]: I1208 14:57:37.907969 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" event={"ID":"383ac257-ce50-4d47-8298-799521c308f6","Type":"ContainerDied","Data":"bbabd7db73c749b6bfbaf48e3f39c48b30f6a6fb6dbedff78c3bf3edadaa1874"} Dec 08 14:57:39 crc kubenswrapper[4894]: I1208 14:57:39.921368 4894 generic.go:334] "Generic (PLEG): container finished" podID="383ac257-ce50-4d47-8298-799521c308f6" containerID="0e1fb35b8a87a4971340a7c552676c248af230d3d94182d608346deeceea9c9e" exitCode=0 Dec 08 14:57:39 crc kubenswrapper[4894]: I1208 14:57:39.921420 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" event={"ID":"383ac257-ce50-4d47-8298-799521c308f6","Type":"ContainerDied","Data":"0e1fb35b8a87a4971340a7c552676c248af230d3d94182d608346deeceea9c9e"} Dec 08 14:57:40 crc kubenswrapper[4894]: I1208 14:57:40.930493 4894 generic.go:334] "Generic (PLEG): container finished" podID="383ac257-ce50-4d47-8298-799521c308f6" containerID="097f89334f33a88b5476efcf7a928d0a390b4fa5d2a868922508ce3c254178ec" exitCode=0 Dec 08 14:57:40 crc kubenswrapper[4894]: I1208 14:57:40.930610 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" event={"ID":"383ac257-ce50-4d47-8298-799521c308f6","Type":"ContainerDied","Data":"097f89334f33a88b5476efcf7a928d0a390b4fa5d2a868922508ce3c254178ec"} Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.157310 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.317727 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-bundle\") pod \"383ac257-ce50-4d47-8298-799521c308f6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.317861 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79djf\" (UniqueName: \"kubernetes.io/projected/383ac257-ce50-4d47-8298-799521c308f6-kube-api-access-79djf\") pod \"383ac257-ce50-4d47-8298-799521c308f6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.317910 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-util\") pod \"383ac257-ce50-4d47-8298-799521c308f6\" (UID: \"383ac257-ce50-4d47-8298-799521c308f6\") " Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.318381 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-bundle" (OuterVolumeSpecName: "bundle") pod "383ac257-ce50-4d47-8298-799521c308f6" (UID: "383ac257-ce50-4d47-8298-799521c308f6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.325874 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/383ac257-ce50-4d47-8298-799521c308f6-kube-api-access-79djf" (OuterVolumeSpecName: "kube-api-access-79djf") pod "383ac257-ce50-4d47-8298-799521c308f6" (UID: "383ac257-ce50-4d47-8298-799521c308f6"). InnerVolumeSpecName "kube-api-access-79djf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.348726 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-util" (OuterVolumeSpecName: "util") pod "383ac257-ce50-4d47-8298-799521c308f6" (UID: "383ac257-ce50-4d47-8298-799521c308f6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.419025 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.419059 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79djf\" (UniqueName: \"kubernetes.io/projected/383ac257-ce50-4d47-8298-799521c308f6-kube-api-access-79djf\") on node \"crc\" DevicePath \"\"" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.419073 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/383ac257-ce50-4d47-8298-799521c308f6-util\") on node \"crc\" DevicePath \"\"" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.946111 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" event={"ID":"383ac257-ce50-4d47-8298-799521c308f6","Type":"ContainerDied","Data":"d6564e0df45946296e04c44a8a8da5d1dced422fcc301ebcfe7206d704884a1f"} Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.946204 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6564e0df45946296e04c44a8a8da5d1dced422fcc301ebcfe7206d704884a1f" Dec 08 14:57:42 crc kubenswrapper[4894]: I1208 14:57:42.946348 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.240358 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb"] Dec 08 14:57:47 crc kubenswrapper[4894]: E1208 14:57:47.241002 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="383ac257-ce50-4d47-8298-799521c308f6" containerName="util" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.241019 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="383ac257-ce50-4d47-8298-799521c308f6" containerName="util" Dec 08 14:57:47 crc kubenswrapper[4894]: E1208 14:57:47.241043 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="383ac257-ce50-4d47-8298-799521c308f6" containerName="extract" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.241054 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="383ac257-ce50-4d47-8298-799521c308f6" containerName="extract" Dec 08 14:57:47 crc kubenswrapper[4894]: E1208 14:57:47.241069 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="383ac257-ce50-4d47-8298-799521c308f6" containerName="pull" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.241079 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="383ac257-ce50-4d47-8298-799521c308f6" containerName="pull" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.241222 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="383ac257-ce50-4d47-8298-799521c308f6" containerName="extract" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.241760 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.244634 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.245279 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mgpbk" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.247229 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.254658 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb"] Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.385283 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7dzj\" (UniqueName: \"kubernetes.io/projected/68045b7d-a426-4739-9870-b4e40c54dedd-kube-api-access-v7dzj\") pod \"nmstate-operator-5b5b58f5c8-7kjmb\" (UID: \"68045b7d-a426-4739-9870-b4e40c54dedd\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.486839 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7dzj\" (UniqueName: \"kubernetes.io/projected/68045b7d-a426-4739-9870-b4e40c54dedd-kube-api-access-v7dzj\") pod \"nmstate-operator-5b5b58f5c8-7kjmb\" (UID: \"68045b7d-a426-4739-9870-b4e40c54dedd\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.511264 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7dzj\" (UniqueName: \"kubernetes.io/projected/68045b7d-a426-4739-9870-b4e40c54dedd-kube-api-access-v7dzj\") pod \"nmstate-operator-5b5b58f5c8-7kjmb\" (UID: \"68045b7d-a426-4739-9870-b4e40c54dedd\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.559078 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.722656 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb"] Dec 08 14:57:47 crc kubenswrapper[4894]: I1208 14:57:47.972736 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" event={"ID":"68045b7d-a426-4739-9870-b4e40c54dedd","Type":"ContainerStarted","Data":"f8a2b3ee1603c11beed7db15da7e2562437b8a63364ef686a47c0d5f100d45dd"} Dec 08 14:57:49 crc kubenswrapper[4894]: I1208 14:57:49.984844 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" event={"ID":"68045b7d-a426-4739-9870-b4e40c54dedd","Type":"ContainerStarted","Data":"772323b84d8b30397b7c35c5a2fb6991e91527a0cffb1e16b035b0c5d0008e36"} Dec 08 14:57:50 crc kubenswrapper[4894]: I1208 14:57:50.008621 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-7kjmb" podStartSLOduration=1.153159495 podStartE2EDuration="3.008601906s" podCreationTimestamp="2025-12-08 14:57:47 +0000 UTC" firstStartedPulling="2025-12-08 14:57:47.731168385 +0000 UTC m=+688.831174500" lastFinishedPulling="2025-12-08 14:57:49.586610796 +0000 UTC m=+690.686616911" observedRunningTime="2025-12-08 14:57:50.004722533 +0000 UTC m=+691.104728668" watchObservedRunningTime="2025-12-08 14:57:50.008601906 +0000 UTC m=+691.108608021" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.353105 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.354437 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.365375 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.366346 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.368084 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.368146 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-rtrsk" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.393680 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.401343 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-b8g67"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.402013 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.418186 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.474532 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.475320 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.476623 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-mxgnm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.479092 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.489878 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.490155 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k8nj\" (UniqueName: \"kubernetes.io/projected/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-kube-api-access-8k8nj\") pod \"nmstate-webhook-5f6d4c5ccb-9wlhf\" (UID: \"70fa62b1-a4b9-49c3-abd2-49ac96c501b6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.490253 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56ddx\" (UniqueName: \"kubernetes.io/projected/2f8534ee-633a-4633-9649-dbc49a705529-kube-api-access-56ddx\") pod \"nmstate-metrics-7f946cbc9-lkw5q\" (UID: \"2f8534ee-633a-4633-9649-dbc49a705529\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.490275 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9wlhf\" (UID: \"70fa62b1-a4b9-49c3-abd2-49ac96c501b6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.493869 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.590980 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-ovs-socket\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591028 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-nmstate-lock\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591059 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2pdc\" (UniqueName: \"kubernetes.io/projected/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-kube-api-access-f2pdc\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591114 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhzgj\" (UniqueName: \"kubernetes.io/projected/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-kube-api-access-hhzgj\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591213 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56ddx\" (UniqueName: \"kubernetes.io/projected/2f8534ee-633a-4633-9649-dbc49a705529-kube-api-access-56ddx\") pod \"nmstate-metrics-7f946cbc9-lkw5q\" (UID: \"2f8534ee-633a-4633-9649-dbc49a705529\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591248 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9wlhf\" (UID: \"70fa62b1-a4b9-49c3-abd2-49ac96c501b6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591271 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k8nj\" (UniqueName: \"kubernetes.io/projected/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-kube-api-access-8k8nj\") pod \"nmstate-webhook-5f6d4c5ccb-9wlhf\" (UID: \"70fa62b1-a4b9-49c3-abd2-49ac96c501b6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591304 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-dbus-socket\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.591352 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: E1208 14:57:56.591721 4894 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 08 14:57:56 crc kubenswrapper[4894]: E1208 14:57:56.591771 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-tls-key-pair podName:70fa62b1-a4b9-49c3-abd2-49ac96c501b6 nodeName:}" failed. No retries permitted until 2025-12-08 14:57:57.091751637 +0000 UTC m=+698.191757752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-9wlhf" (UID: "70fa62b1-a4b9-49c3-abd2-49ac96c501b6") : secret "openshift-nmstate-webhook" not found Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.623731 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56ddx\" (UniqueName: \"kubernetes.io/projected/2f8534ee-633a-4633-9649-dbc49a705529-kube-api-access-56ddx\") pod \"nmstate-metrics-7f946cbc9-lkw5q\" (UID: \"2f8534ee-633a-4633-9649-dbc49a705529\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.623735 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k8nj\" (UniqueName: \"kubernetes.io/projected/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-kube-api-access-8k8nj\") pod \"nmstate-webhook-5f6d4c5ccb-9wlhf\" (UID: \"70fa62b1-a4b9-49c3-abd2-49ac96c501b6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.675759 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.677629 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5698ddd759-ztbcm"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.678502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.689301 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5698ddd759-ztbcm"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.693444 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-dbus-socket\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.693523 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.693566 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-ovs-socket\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.693590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-nmstate-lock\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.693616 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2pdc\" (UniqueName: \"kubernetes.io/projected/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-kube-api-access-f2pdc\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.693635 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.693665 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhzgj\" (UniqueName: \"kubernetes.io/projected/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-kube-api-access-hhzgj\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.694249 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-dbus-socket\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.696919 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-nmstate-lock\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.696995 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-ovs-socket\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.698215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.703473 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.722924 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2pdc\" (UniqueName: \"kubernetes.io/projected/55c7d1cf-5cb4-4a06-ac50-2ee06f72a522-kube-api-access-f2pdc\") pod \"nmstate-console-plugin-7fbb5f6569-5v49c\" (UID: \"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.723624 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhzgj\" (UniqueName: \"kubernetes.io/projected/2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c-kube-api-access-hhzgj\") pod \"nmstate-handler-b8g67\" (UID: \"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c\") " pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.793571 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.794425 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8011f339-5239-4484-bf55-5e8661038612-console-oauth-config\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.794464 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-trusted-ca-bundle\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.794487 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8011f339-5239-4484-bf55-5e8661038612-console-serving-cert\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.794515 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-service-ca\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.794557 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-oauth-serving-cert\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.794587 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-console-config\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.794658 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69gw2\" (UniqueName: \"kubernetes.io/projected/8011f339-5239-4484-bf55-5e8661038612-kube-api-access-69gw2\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.891320 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q"] Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.895542 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69gw2\" (UniqueName: \"kubernetes.io/projected/8011f339-5239-4484-bf55-5e8661038612-kube-api-access-69gw2\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.895593 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8011f339-5239-4484-bf55-5e8661038612-console-oauth-config\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.895619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-trusted-ca-bundle\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.895641 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8011f339-5239-4484-bf55-5e8661038612-console-serving-cert\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.895668 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-service-ca\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.895698 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-oauth-serving-cert\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.895729 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-console-config\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.896777 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-console-config\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.896977 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-oauth-serving-cert\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.897076 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-service-ca\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.897864 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8011f339-5239-4484-bf55-5e8661038612-trusted-ca-bundle\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.899782 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8011f339-5239-4484-bf55-5e8661038612-console-serving-cert\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.901221 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8011f339-5239-4484-bf55-5e8661038612-console-oauth-config\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.912506 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69gw2\" (UniqueName: \"kubernetes.io/projected/8011f339-5239-4484-bf55-5e8661038612-kube-api-access-69gw2\") pod \"console-5698ddd759-ztbcm\" (UID: \"8011f339-5239-4484-bf55-5e8661038612\") " pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:56 crc kubenswrapper[4894]: I1208 14:57:56.972793 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c"] Dec 08 14:57:56 crc kubenswrapper[4894]: W1208 14:57:56.977677 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55c7d1cf_5cb4_4a06_ac50_2ee06f72a522.slice/crio-07396ca619a416c5adf64104d7bd0777dc3fc0f225b3a512e507242fe10c8119 WatchSource:0}: Error finding container 07396ca619a416c5adf64104d7bd0777dc3fc0f225b3a512e507242fe10c8119: Status 404 returned error can't find the container with id 07396ca619a416c5adf64104d7bd0777dc3fc0f225b3a512e507242fe10c8119 Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.017682 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.022278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" event={"ID":"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522","Type":"ContainerStarted","Data":"07396ca619a416c5adf64104d7bd0777dc3fc0f225b3a512e507242fe10c8119"} Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.023495 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" event={"ID":"2f8534ee-633a-4633-9649-dbc49a705529","Type":"ContainerStarted","Data":"6293863056a534a0778ae39fa88173d3e7fde0ec729f7aaa74662176eadbf6ba"} Dec 08 14:57:57 crc kubenswrapper[4894]: W1208 14:57:57.038794 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ed9116e_64a3_4bc3_a4fc_50e2a7fd383c.slice/crio-09af59fd0575c7bec301483a0b80547b4eecc9d643f0c661ba2c090cb7122d6d WatchSource:0}: Error finding container 09af59fd0575c7bec301483a0b80547b4eecc9d643f0c661ba2c090cb7122d6d: Status 404 returned error can't find the container with id 09af59fd0575c7bec301483a0b80547b4eecc9d643f0c661ba2c090cb7122d6d Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.055628 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.097565 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9wlhf\" (UID: \"70fa62b1-a4b9-49c3-abd2-49ac96c501b6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.100689 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/70fa62b1-a4b9-49c3-abd2-49ac96c501b6-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9wlhf\" (UID: \"70fa62b1-a4b9-49c3-abd2-49ac96c501b6\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.241159 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5698ddd759-ztbcm"] Dec 08 14:57:57 crc kubenswrapper[4894]: W1208 14:57:57.247651 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8011f339_5239_4484_bf55_5e8661038612.slice/crio-dd31f6ba828139b9ff26458863a23f6d886d945c6b382ef2f9a8e282da8709d4 WatchSource:0}: Error finding container dd31f6ba828139b9ff26458863a23f6d886d945c6b382ef2f9a8e282da8709d4: Status 404 returned error can't find the container with id dd31f6ba828139b9ff26458863a23f6d886d945c6b382ef2f9a8e282da8709d4 Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.290477 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:57:57 crc kubenswrapper[4894]: I1208 14:57:57.466922 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf"] Dec 08 14:57:57 crc kubenswrapper[4894]: W1208 14:57:57.472073 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70fa62b1_a4b9_49c3_abd2_49ac96c501b6.slice/crio-f1d102067983e197d4a6e042b7149f97cc4dea25e1358dec31a4e9153acdb171 WatchSource:0}: Error finding container f1d102067983e197d4a6e042b7149f97cc4dea25e1358dec31a4e9153acdb171: Status 404 returned error can't find the container with id f1d102067983e197d4a6e042b7149f97cc4dea25e1358dec31a4e9153acdb171 Dec 08 14:57:58 crc kubenswrapper[4894]: I1208 14:57:58.033293 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5698ddd759-ztbcm" event={"ID":"8011f339-5239-4484-bf55-5e8661038612","Type":"ContainerStarted","Data":"fc9c78533e7e0005c279fedf86babfd12451a321bfc9ecc73a7e827af04c23ba"} Dec 08 14:57:58 crc kubenswrapper[4894]: I1208 14:57:58.033617 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5698ddd759-ztbcm" event={"ID":"8011f339-5239-4484-bf55-5e8661038612","Type":"ContainerStarted","Data":"dd31f6ba828139b9ff26458863a23f6d886d945c6b382ef2f9a8e282da8709d4"} Dec 08 14:57:58 crc kubenswrapper[4894]: I1208 14:57:58.036869 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b8g67" event={"ID":"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c","Type":"ContainerStarted","Data":"09af59fd0575c7bec301483a0b80547b4eecc9d643f0c661ba2c090cb7122d6d"} Dec 08 14:57:58 crc kubenswrapper[4894]: I1208 14:57:58.037809 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" event={"ID":"70fa62b1-a4b9-49c3-abd2-49ac96c501b6","Type":"ContainerStarted","Data":"f1d102067983e197d4a6e042b7149f97cc4dea25e1358dec31a4e9153acdb171"} Dec 08 14:57:58 crc kubenswrapper[4894]: I1208 14:57:58.053649 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5698ddd759-ztbcm" podStartSLOduration=2.053627028 podStartE2EDuration="2.053627028s" podCreationTimestamp="2025-12-08 14:57:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:57:58.048587419 +0000 UTC m=+699.148593554" watchObservedRunningTime="2025-12-08 14:57:58.053627028 +0000 UTC m=+699.153633143" Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.048271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" event={"ID":"70fa62b1-a4b9-49c3-abd2-49ac96c501b6","Type":"ContainerStarted","Data":"9b68191d797d08b4690c8339b58939793c2f86ed40e7140ea4f042f73190b795"} Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.048906 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.050117 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b8g67" event={"ID":"2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c","Type":"ContainerStarted","Data":"b029488dcdaa974cc095aa7db17f9183aff0aaab932a988d41d3c09f53584513"} Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.050195 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.053273 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" event={"ID":"2f8534ee-633a-4633-9649-dbc49a705529","Type":"ContainerStarted","Data":"ea831ba2198e4b653f9ecd85b54b2284919f3cb217bf67c7a94c606e6f2d675f"} Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.054637 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" event={"ID":"55c7d1cf-5cb4-4a06-ac50-2ee06f72a522","Type":"ContainerStarted","Data":"81ae3c9c3f85d498936886b6ca59646dd671089bb17d3e64abc4fab1fb4f31e0"} Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.067052 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" podStartSLOduration=2.335101033 podStartE2EDuration="4.067027017s" podCreationTimestamp="2025-12-08 14:57:56 +0000 UTC" firstStartedPulling="2025-12-08 14:57:57.474801646 +0000 UTC m=+698.574807761" lastFinishedPulling="2025-12-08 14:57:59.20672763 +0000 UTC m=+700.306733745" observedRunningTime="2025-12-08 14:58:00.061178892 +0000 UTC m=+701.161185017" watchObservedRunningTime="2025-12-08 14:58:00.067027017 +0000 UTC m=+701.167033132" Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.080698 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-b8g67" podStartSLOduration=1.914327703 podStartE2EDuration="4.08067576s" podCreationTimestamp="2025-12-08 14:57:56 +0000 UTC" firstStartedPulling="2025-12-08 14:57:57.041219429 +0000 UTC m=+698.141225544" lastFinishedPulling="2025-12-08 14:57:59.207567486 +0000 UTC m=+700.307573601" observedRunningTime="2025-12-08 14:58:00.076517639 +0000 UTC m=+701.176523764" watchObservedRunningTime="2025-12-08 14:58:00.08067576 +0000 UTC m=+701.180681885" Dec 08 14:58:00 crc kubenswrapper[4894]: I1208 14:58:00.099854 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5v49c" podStartSLOduration=1.881759738 podStartE2EDuration="4.099837557s" podCreationTimestamp="2025-12-08 14:57:56 +0000 UTC" firstStartedPulling="2025-12-08 14:57:56.979611155 +0000 UTC m=+698.079617270" lastFinishedPulling="2025-12-08 14:57:59.197688974 +0000 UTC m=+700.297695089" observedRunningTime="2025-12-08 14:58:00.092391512 +0000 UTC m=+701.192397627" watchObservedRunningTime="2025-12-08 14:58:00.099837557 +0000 UTC m=+701.199843672" Dec 08 14:58:02 crc kubenswrapper[4894]: I1208 14:58:02.073286 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" event={"ID":"2f8534ee-633a-4633-9649-dbc49a705529","Type":"ContainerStarted","Data":"52175bfa433f6e5045ae395737fce7e93fd299427045835afb8fcb035867b921"} Dec 08 14:58:02 crc kubenswrapper[4894]: I1208 14:58:02.088289 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-lkw5q" podStartSLOduration=1.805532841 podStartE2EDuration="6.088235383s" podCreationTimestamp="2025-12-08 14:57:56 +0000 UTC" firstStartedPulling="2025-12-08 14:57:56.902025705 +0000 UTC m=+698.002031820" lastFinishedPulling="2025-12-08 14:58:01.184728247 +0000 UTC m=+702.284734362" observedRunningTime="2025-12-08 14:58:02.086205789 +0000 UTC m=+703.186211924" watchObservedRunningTime="2025-12-08 14:58:02.088235383 +0000 UTC m=+703.188241498" Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.037706 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-b8g67" Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.056460 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.058116 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.064231 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.108331 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5698ddd759-ztbcm" Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.187303 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-77css"] Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.296428 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:58:07 crc kubenswrapper[4894]: I1208 14:58:07.296727 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:58:17 crc kubenswrapper[4894]: I1208 14:58:17.296372 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9wlhf" Dec 08 14:58:29 crc kubenswrapper[4894]: I1208 14:58:29.861534 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x"] Dec 08 14:58:29 crc kubenswrapper[4894]: I1208 14:58:29.863245 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:29 crc kubenswrapper[4894]: I1208 14:58:29.865137 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 08 14:58:29 crc kubenswrapper[4894]: I1208 14:58:29.870446 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x"] Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.023262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.023355 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r4mb\" (UniqueName: \"kubernetes.io/projected/f40ea837-9c2e-4693-960d-246a8ad84e30-kube-api-access-7r4mb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.023419 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.124945 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.125010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r4mb\" (UniqueName: \"kubernetes.io/projected/f40ea837-9c2e-4693-960d-246a8ad84e30-kube-api-access-7r4mb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.125050 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.125470 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.125515 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.143414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r4mb\" (UniqueName: \"kubernetes.io/projected/f40ea837-9c2e-4693-960d-246a8ad84e30-kube-api-access-7r4mb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.181032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:30 crc kubenswrapper[4894]: I1208 14:58:30.565378 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x"] Dec 08 14:58:31 crc kubenswrapper[4894]: I1208 14:58:31.227907 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" event={"ID":"f40ea837-9c2e-4693-960d-246a8ad84e30","Type":"ContainerStarted","Data":"dda54fc4b04b40a629ed2bb22f90200331def6cd5bba62bca3b434e8da8cc334"} Dec 08 14:58:31 crc kubenswrapper[4894]: I1208 14:58:31.227949 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" event={"ID":"f40ea837-9c2e-4693-960d-246a8ad84e30","Type":"ContainerStarted","Data":"640c08cbe73aa14391d431848250a4ae0e67b4d19afe94f2bbd028e6a302994d"} Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.236457 4894 generic.go:334] "Generic (PLEG): container finished" podID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerID="dda54fc4b04b40a629ed2bb22f90200331def6cd5bba62bca3b434e8da8cc334" exitCode=0 Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.236500 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" event={"ID":"f40ea837-9c2e-4693-960d-246a8ad84e30","Type":"ContainerDied","Data":"dda54fc4b04b40a629ed2bb22f90200331def6cd5bba62bca3b434e8da8cc334"} Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.237583 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-77css" podUID="368e61a3-8284-46f3-8901-cdcad7a729cb" containerName="console" containerID="cri-o://a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac" gracePeriod=15 Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.597227 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-77css_368e61a3-8284-46f3-8901-cdcad7a729cb/console/0.log" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.597290 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.762124 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-serving-cert\") pod \"368e61a3-8284-46f3-8901-cdcad7a729cb\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.762239 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97gr2\" (UniqueName: \"kubernetes.io/projected/368e61a3-8284-46f3-8901-cdcad7a729cb-kube-api-access-97gr2\") pod \"368e61a3-8284-46f3-8901-cdcad7a729cb\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.762295 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-service-ca\") pod \"368e61a3-8284-46f3-8901-cdcad7a729cb\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.762346 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-console-config\") pod \"368e61a3-8284-46f3-8901-cdcad7a729cb\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.762378 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-oauth-config\") pod \"368e61a3-8284-46f3-8901-cdcad7a729cb\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.762445 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-trusted-ca-bundle\") pod \"368e61a3-8284-46f3-8901-cdcad7a729cb\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.762476 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-oauth-serving-cert\") pod \"368e61a3-8284-46f3-8901-cdcad7a729cb\" (UID: \"368e61a3-8284-46f3-8901-cdcad7a729cb\") " Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.763589 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "368e61a3-8284-46f3-8901-cdcad7a729cb" (UID: "368e61a3-8284-46f3-8901-cdcad7a729cb"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.763669 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "368e61a3-8284-46f3-8901-cdcad7a729cb" (UID: "368e61a3-8284-46f3-8901-cdcad7a729cb"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.763764 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-console-config" (OuterVolumeSpecName: "console-config") pod "368e61a3-8284-46f3-8901-cdcad7a729cb" (UID: "368e61a3-8284-46f3-8901-cdcad7a729cb"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.764045 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-service-ca" (OuterVolumeSpecName: "service-ca") pod "368e61a3-8284-46f3-8901-cdcad7a729cb" (UID: "368e61a3-8284-46f3-8901-cdcad7a729cb"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.768754 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "368e61a3-8284-46f3-8901-cdcad7a729cb" (UID: "368e61a3-8284-46f3-8901-cdcad7a729cb"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.770902 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/368e61a3-8284-46f3-8901-cdcad7a729cb-kube-api-access-97gr2" (OuterVolumeSpecName: "kube-api-access-97gr2") pod "368e61a3-8284-46f3-8901-cdcad7a729cb" (UID: "368e61a3-8284-46f3-8901-cdcad7a729cb"). InnerVolumeSpecName "kube-api-access-97gr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.771022 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "368e61a3-8284-46f3-8901-cdcad7a729cb" (UID: "368e61a3-8284-46f3-8901-cdcad7a729cb"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.864498 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97gr2\" (UniqueName: \"kubernetes.io/projected/368e61a3-8284-46f3-8901-cdcad7a729cb-kube-api-access-97gr2\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.864550 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-service-ca\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.864565 4894 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-console-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.864576 4894 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.864588 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.864599 4894 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/368e61a3-8284-46f3-8901-cdcad7a729cb-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:32 crc kubenswrapper[4894]: I1208 14:58:32.864611 4894 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/368e61a3-8284-46f3-8901-cdcad7a729cb-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.242366 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-77css_368e61a3-8284-46f3-8901-cdcad7a729cb/console/0.log" Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.242411 4894 generic.go:334] "Generic (PLEG): container finished" podID="368e61a3-8284-46f3-8901-cdcad7a729cb" containerID="a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac" exitCode=2 Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.242440 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-77css" event={"ID":"368e61a3-8284-46f3-8901-cdcad7a729cb","Type":"ContainerDied","Data":"a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac"} Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.242467 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-77css" event={"ID":"368e61a3-8284-46f3-8901-cdcad7a729cb","Type":"ContainerDied","Data":"48d9156706719c43dbfc8101ade6556d42879d64a082646b371f10bfedf625f3"} Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.242486 4894 scope.go:117] "RemoveContainer" containerID="a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac" Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.242600 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-77css" Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.263214 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-77css"] Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.267949 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-77css"] Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.312158 4894 scope.go:117] "RemoveContainer" containerID="a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac" Dec 08 14:58:33 crc kubenswrapper[4894]: E1208 14:58:33.312685 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac\": container with ID starting with a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac not found: ID does not exist" containerID="a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac" Dec 08 14:58:33 crc kubenswrapper[4894]: I1208 14:58:33.312883 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac"} err="failed to get container status \"a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac\": rpc error: code = NotFound desc = could not find container \"a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac\": container with ID starting with a43af64198d505a9df2f947b358b59a8ed459c06411ca9baa06697ffbb8783ac not found: ID does not exist" Dec 08 14:58:34 crc kubenswrapper[4894]: I1208 14:58:34.249851 4894 generic.go:334] "Generic (PLEG): container finished" podID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerID="8641ebc715e0503c48912eb68a5e2bf088ee82a030d8128c6950da607b0bb982" exitCode=0 Dec 08 14:58:34 crc kubenswrapper[4894]: I1208 14:58:34.249912 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" event={"ID":"f40ea837-9c2e-4693-960d-246a8ad84e30","Type":"ContainerDied","Data":"8641ebc715e0503c48912eb68a5e2bf088ee82a030d8128c6950da607b0bb982"} Dec 08 14:58:35 crc kubenswrapper[4894]: I1208 14:58:35.207255 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="368e61a3-8284-46f3-8901-cdcad7a729cb" path="/var/lib/kubelet/pods/368e61a3-8284-46f3-8901-cdcad7a729cb/volumes" Dec 08 14:58:35 crc kubenswrapper[4894]: I1208 14:58:35.258611 4894 generic.go:334] "Generic (PLEG): container finished" podID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerID="db358382cc7542eaf12db543945253c0c0e93fa7292ecfa23f334016146b6540" exitCode=0 Dec 08 14:58:35 crc kubenswrapper[4894]: I1208 14:58:35.258662 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" event={"ID":"f40ea837-9c2e-4693-960d-246a8ad84e30","Type":"ContainerDied","Data":"db358382cc7542eaf12db543945253c0c0e93fa7292ecfa23f334016146b6540"} Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.508610 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.609401 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-util\") pod \"f40ea837-9c2e-4693-960d-246a8ad84e30\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.609444 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r4mb\" (UniqueName: \"kubernetes.io/projected/f40ea837-9c2e-4693-960d-246a8ad84e30-kube-api-access-7r4mb\") pod \"f40ea837-9c2e-4693-960d-246a8ad84e30\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.609558 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-bundle\") pod \"f40ea837-9c2e-4693-960d-246a8ad84e30\" (UID: \"f40ea837-9c2e-4693-960d-246a8ad84e30\") " Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.610841 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-bundle" (OuterVolumeSpecName: "bundle") pod "f40ea837-9c2e-4693-960d-246a8ad84e30" (UID: "f40ea837-9c2e-4693-960d-246a8ad84e30"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.614508 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40ea837-9c2e-4693-960d-246a8ad84e30-kube-api-access-7r4mb" (OuterVolumeSpecName: "kube-api-access-7r4mb") pod "f40ea837-9c2e-4693-960d-246a8ad84e30" (UID: "f40ea837-9c2e-4693-960d-246a8ad84e30"). InnerVolumeSpecName "kube-api-access-7r4mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.625104 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-util" (OuterVolumeSpecName: "util") pod "f40ea837-9c2e-4693-960d-246a8ad84e30" (UID: "f40ea837-9c2e-4693-960d-246a8ad84e30"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.711602 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-util\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.711645 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r4mb\" (UniqueName: \"kubernetes.io/projected/f40ea837-9c2e-4693-960d-246a8ad84e30-kube-api-access-7r4mb\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:36 crc kubenswrapper[4894]: I1208 14:58:36.711657 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f40ea837-9c2e-4693-960d-246a8ad84e30-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 14:58:37 crc kubenswrapper[4894]: I1208 14:58:37.270056 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" event={"ID":"f40ea837-9c2e-4693-960d-246a8ad84e30","Type":"ContainerDied","Data":"640c08cbe73aa14391d431848250a4ae0e67b4d19afe94f2bbd028e6a302994d"} Dec 08 14:58:37 crc kubenswrapper[4894]: I1208 14:58:37.270369 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="640c08cbe73aa14391d431848250a4ae0e67b4d19afe94f2bbd028e6a302994d" Dec 08 14:58:37 crc kubenswrapper[4894]: I1208 14:58:37.270176 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x" Dec 08 14:58:37 crc kubenswrapper[4894]: I1208 14:58:37.297256 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:58:37 crc kubenswrapper[4894]: I1208 14:58:37.297310 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:58:47 crc kubenswrapper[4894]: I1208 14:58:47.421705 4894 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.790178 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-78884d4998-whjq5"] Dec 08 14:58:50 crc kubenswrapper[4894]: E1208 14:58:50.790648 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerName="extract" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.790660 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerName="extract" Dec 08 14:58:50 crc kubenswrapper[4894]: E1208 14:58:50.790667 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="368e61a3-8284-46f3-8901-cdcad7a729cb" containerName="console" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.790673 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="368e61a3-8284-46f3-8901-cdcad7a729cb" containerName="console" Dec 08 14:58:50 crc kubenswrapper[4894]: E1208 14:58:50.790688 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerName="util" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.790694 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerName="util" Dec 08 14:58:50 crc kubenswrapper[4894]: E1208 14:58:50.790702 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerName="pull" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.790708 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerName="pull" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.790795 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40ea837-9c2e-4693-960d-246a8ad84e30" containerName="extract" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.790804 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="368e61a3-8284-46f3-8901-cdcad7a729cb" containerName="console" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.791150 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.803114 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-pvfn2" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.803340 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.803443 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.803124 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.814787 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.818845 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78884d4998-whjq5"] Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.880314 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfr59\" (UniqueName: \"kubernetes.io/projected/10bcfd29-cff4-4fcc-be38-43296e67cac7-kube-api-access-sfr59\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.880378 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10bcfd29-cff4-4fcc-be38-43296e67cac7-webhook-cert\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.880405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10bcfd29-cff4-4fcc-be38-43296e67cac7-apiservice-cert\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.981847 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfr59\" (UniqueName: \"kubernetes.io/projected/10bcfd29-cff4-4fcc-be38-43296e67cac7-kube-api-access-sfr59\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.981918 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10bcfd29-cff4-4fcc-be38-43296e67cac7-webhook-cert\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.981948 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10bcfd29-cff4-4fcc-be38-43296e67cac7-apiservice-cert\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.989134 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10bcfd29-cff4-4fcc-be38-43296e67cac7-apiservice-cert\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:50 crc kubenswrapper[4894]: I1208 14:58:50.994968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10bcfd29-cff4-4fcc-be38-43296e67cac7-webhook-cert\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.007545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfr59\" (UniqueName: \"kubernetes.io/projected/10bcfd29-cff4-4fcc-be38-43296e67cac7-kube-api-access-sfr59\") pod \"metallb-operator-controller-manager-78884d4998-whjq5\" (UID: \"10bcfd29-cff4-4fcc-be38-43296e67cac7\") " pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.040798 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6"] Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.041994 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.043439 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-kzt9p" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.043720 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.054247 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.072892 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6"] Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.120081 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.184189 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-webhook-cert\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.184270 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2rx5\" (UniqueName: \"kubernetes.io/projected/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-kube-api-access-w2rx5\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.184298 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-apiservice-cert\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.287276 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-webhook-cert\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.287350 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2rx5\" (UniqueName: \"kubernetes.io/projected/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-kube-api-access-w2rx5\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.287378 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-apiservice-cert\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.291866 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-apiservice-cert\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.311162 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-webhook-cert\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.312194 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2rx5\" (UniqueName: \"kubernetes.io/projected/6e5ec2a2-054b-4cb6-a3d6-7376f254869c-kube-api-access-w2rx5\") pod \"metallb-operator-webhook-server-74cdf75695-th2w6\" (UID: \"6e5ec2a2-054b-4cb6-a3d6-7376f254869c\") " pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.360556 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.453347 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78884d4998-whjq5"] Dec 08 14:58:51 crc kubenswrapper[4894]: W1208 14:58:51.470504 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10bcfd29_cff4_4fcc_be38_43296e67cac7.slice/crio-bafa24ed75e06375b900b09ae7e1a564a219ef382ec583eee26cd6276bc5309b WatchSource:0}: Error finding container bafa24ed75e06375b900b09ae7e1a564a219ef382ec583eee26cd6276bc5309b: Status 404 returned error can't find the container with id bafa24ed75e06375b900b09ae7e1a564a219ef382ec583eee26cd6276bc5309b Dec 08 14:58:51 crc kubenswrapper[4894]: I1208 14:58:51.817919 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6"] Dec 08 14:58:51 crc kubenswrapper[4894]: W1208 14:58:51.823490 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e5ec2a2_054b_4cb6_a3d6_7376f254869c.slice/crio-d632f18f939981c5e1c504d5e6ef94ad3a54b0fe43fc6e196470e8c0203a8708 WatchSource:0}: Error finding container d632f18f939981c5e1c504d5e6ef94ad3a54b0fe43fc6e196470e8c0203a8708: Status 404 returned error can't find the container with id d632f18f939981c5e1c504d5e6ef94ad3a54b0fe43fc6e196470e8c0203a8708 Dec 08 14:58:52 crc kubenswrapper[4894]: I1208 14:58:52.354533 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" event={"ID":"6e5ec2a2-054b-4cb6-a3d6-7376f254869c","Type":"ContainerStarted","Data":"d632f18f939981c5e1c504d5e6ef94ad3a54b0fe43fc6e196470e8c0203a8708"} Dec 08 14:58:52 crc kubenswrapper[4894]: I1208 14:58:52.355295 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" event={"ID":"10bcfd29-cff4-4fcc-be38-43296e67cac7","Type":"ContainerStarted","Data":"bafa24ed75e06375b900b09ae7e1a564a219ef382ec583eee26cd6276bc5309b"} Dec 08 14:58:55 crc kubenswrapper[4894]: I1208 14:58:55.374339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" event={"ID":"10bcfd29-cff4-4fcc-be38-43296e67cac7","Type":"ContainerStarted","Data":"f426d13c24f7b10fbceeaade1081aa9b2f1ecf48a7cf89e589a03d9a7c88b283"} Dec 08 14:58:55 crc kubenswrapper[4894]: I1208 14:58:55.374976 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:58:55 crc kubenswrapper[4894]: I1208 14:58:55.409135 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" podStartSLOduration=2.31638545 podStartE2EDuration="5.409113923s" podCreationTimestamp="2025-12-08 14:58:50 +0000 UTC" firstStartedPulling="2025-12-08 14:58:51.473543359 +0000 UTC m=+752.573549474" lastFinishedPulling="2025-12-08 14:58:54.566271832 +0000 UTC m=+755.666277947" observedRunningTime="2025-12-08 14:58:55.404354914 +0000 UTC m=+756.504361039" watchObservedRunningTime="2025-12-08 14:58:55.409113923 +0000 UTC m=+756.509120048" Dec 08 14:58:57 crc kubenswrapper[4894]: I1208 14:58:57.394290 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" event={"ID":"6e5ec2a2-054b-4cb6-a3d6-7376f254869c","Type":"ContainerStarted","Data":"94d1610f9dc6b9769ec96793c4563c110f3fa2257aa6a08c4280dc5d3a278b45"} Dec 08 14:58:57 crc kubenswrapper[4894]: I1208 14:58:57.394918 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:58:57 crc kubenswrapper[4894]: I1208 14:58:57.416547 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" podStartSLOduration=1.415008439 podStartE2EDuration="6.416517461s" podCreationTimestamp="2025-12-08 14:58:51 +0000 UTC" firstStartedPulling="2025-12-08 14:58:51.82634354 +0000 UTC m=+752.926349655" lastFinishedPulling="2025-12-08 14:58:56.827852562 +0000 UTC m=+757.927858677" observedRunningTime="2025-12-08 14:58:57.412081393 +0000 UTC m=+758.512087518" watchObservedRunningTime="2025-12-08 14:58:57.416517461 +0000 UTC m=+758.516523616" Dec 08 14:59:07 crc kubenswrapper[4894]: I1208 14:59:07.296842 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 14:59:07 crc kubenswrapper[4894]: I1208 14:59:07.297380 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 14:59:07 crc kubenswrapper[4894]: I1208 14:59:07.297429 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 14:59:07 crc kubenswrapper[4894]: I1208 14:59:07.298040 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfc08546d7b8c2b8b1aa3ec04284ba529dbbd4025e73389fcf8cb4e001930019"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 14:59:07 crc kubenswrapper[4894]: I1208 14:59:07.298103 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://dfc08546d7b8c2b8b1aa3ec04284ba529dbbd4025e73389fcf8cb4e001930019" gracePeriod=600 Dec 08 14:59:08 crc kubenswrapper[4894]: I1208 14:59:08.451428 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="dfc08546d7b8c2b8b1aa3ec04284ba529dbbd4025e73389fcf8cb4e001930019" exitCode=0 Dec 08 14:59:08 crc kubenswrapper[4894]: I1208 14:59:08.451496 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"dfc08546d7b8c2b8b1aa3ec04284ba529dbbd4025e73389fcf8cb4e001930019"} Dec 08 14:59:08 crc kubenswrapper[4894]: I1208 14:59:08.451990 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"1b3439967813e9136ef0135406ae68191aab690f31fec7c75b03962a836112d1"} Dec 08 14:59:08 crc kubenswrapper[4894]: I1208 14:59:08.452012 4894 scope.go:117] "RemoveContainer" containerID="bcf05c5956a90ca1ef0b1249396e38947db07b49d45785534a1f514e4747c039" Dec 08 14:59:11 crc kubenswrapper[4894]: I1208 14:59:11.365878 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-74cdf75695-th2w6" Dec 08 14:59:31 crc kubenswrapper[4894]: I1208 14:59:31.124055 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-78884d4998-whjq5" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.072665 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-5r5t2"] Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.075689 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.077403 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4"] Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.078000 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.078160 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6qgl7" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.078199 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.078853 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.081129 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.087676 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4"] Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.155464 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hxwz6"] Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.156950 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.159364 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.159416 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.159839 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.160359 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qlc68" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.180495 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-6m7rl"] Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.181422 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.189612 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.209372 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-frr-sockets\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.212534 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-reloader\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.212611 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcmbz\" (UniqueName: \"kubernetes.io/projected/29388ab6-9e49-4662-8d74-1dda6e8d228e-kube-api-access-jcmbz\") pod \"frr-k8s-webhook-server-7fcb986d4-rt5h4\" (UID: \"29388ab6-9e49-4662-8d74-1dda6e8d228e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.212634 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e7389153-ceef-48e4-a814-b6204ce84627-frr-startup\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.212701 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7389153-ceef-48e4-a814-b6204ce84627-metrics-certs\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.218088 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-frr-conf\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.218125 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29388ab6-9e49-4662-8d74-1dda6e8d228e-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rt5h4\" (UID: \"29388ab6-9e49-4662-8d74-1dda6e8d228e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.218163 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbrsr\" (UniqueName: \"kubernetes.io/projected/e7389153-ceef-48e4-a814-b6204ce84627-kube-api-access-lbrsr\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.218191 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-metrics\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.222941 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6m7rl"] Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321298 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf8mg\" (UniqueName: \"kubernetes.io/projected/e316690b-6982-4588-ab38-a33ca86ba1f0-kube-api-access-hf8mg\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2pcb\" (UniqueName: \"kubernetes.io/projected/e2473db3-776f-4a3d-9b4f-aba30d96463b-kube-api-access-k2pcb\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321371 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7389153-ceef-48e4-a814-b6204ce84627-metrics-certs\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2473db3-776f-4a3d-9b4f-aba30d96463b-metrics-certs\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321436 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-metrics-certs\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321460 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e316690b-6982-4588-ab38-a33ca86ba1f0-metallb-excludel2\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321489 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-memberlist\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321674 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-frr-conf\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321755 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29388ab6-9e49-4662-8d74-1dda6e8d228e-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rt5h4\" (UID: \"29388ab6-9e49-4662-8d74-1dda6e8d228e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbrsr\" (UniqueName: \"kubernetes.io/projected/e7389153-ceef-48e4-a814-b6204ce84627-kube-api-access-lbrsr\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321868 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-metrics\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321955 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-frr-sockets\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321976 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-reloader\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: E1208 14:59:32.321983 4894 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 08 14:59:32 crc kubenswrapper[4894]: E1208 14:59:32.322061 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29388ab6-9e49-4662-8d74-1dda6e8d228e-cert podName:29388ab6-9e49-4662-8d74-1dda6e8d228e nodeName:}" failed. No retries permitted until 2025-12-08 14:59:32.822040818 +0000 UTC m=+793.922046933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29388ab6-9e49-4662-8d74-1dda6e8d228e-cert") pod "frr-k8s-webhook-server-7fcb986d4-rt5h4" (UID: "29388ab6-9e49-4662-8d74-1dda6e8d228e") : secret "frr-k8s-webhook-server-cert" not found Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.321991 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e7389153-ceef-48e4-a814-b6204ce84627-frr-startup\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.322120 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-frr-conf\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.322179 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcmbz\" (UniqueName: \"kubernetes.io/projected/29388ab6-9e49-4662-8d74-1dda6e8d228e-kube-api-access-jcmbz\") pod \"frr-k8s-webhook-server-7fcb986d4-rt5h4\" (UID: \"29388ab6-9e49-4662-8d74-1dda6e8d228e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.322430 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-metrics\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.322668 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-frr-sockets\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.322713 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e7389153-ceef-48e4-a814-b6204ce84627-reloader\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.322731 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2473db3-776f-4a3d-9b4f-aba30d96463b-cert\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.322943 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e7389153-ceef-48e4-a814-b6204ce84627-frr-startup\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.327306 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e7389153-ceef-48e4-a814-b6204ce84627-metrics-certs\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.339394 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbrsr\" (UniqueName: \"kubernetes.io/projected/e7389153-ceef-48e4-a814-b6204ce84627-kube-api-access-lbrsr\") pod \"frr-k8s-5r5t2\" (UID: \"e7389153-ceef-48e4-a814-b6204ce84627\") " pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.354211 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcmbz\" (UniqueName: \"kubernetes.io/projected/29388ab6-9e49-4662-8d74-1dda6e8d228e-kube-api-access-jcmbz\") pod \"frr-k8s-webhook-server-7fcb986d4-rt5h4\" (UID: \"29388ab6-9e49-4662-8d74-1dda6e8d228e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.400540 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.423518 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf8mg\" (UniqueName: \"kubernetes.io/projected/e316690b-6982-4588-ab38-a33ca86ba1f0-kube-api-access-hf8mg\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.423569 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2pcb\" (UniqueName: \"kubernetes.io/projected/e2473db3-776f-4a3d-9b4f-aba30d96463b-kube-api-access-k2pcb\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.423619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2473db3-776f-4a3d-9b4f-aba30d96463b-metrics-certs\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.423647 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-metrics-certs\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.423664 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e316690b-6982-4588-ab38-a33ca86ba1f0-metallb-excludel2\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.423684 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-memberlist\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.423745 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2473db3-776f-4a3d-9b4f-aba30d96463b-cert\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.425522 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e316690b-6982-4588-ab38-a33ca86ba1f0-metallb-excludel2\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: E1208 14:59:32.427623 4894 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 08 14:59:32 crc kubenswrapper[4894]: E1208 14:59:32.427709 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-memberlist podName:e316690b-6982-4588-ab38-a33ca86ba1f0 nodeName:}" failed. No retries permitted until 2025-12-08 14:59:32.927686687 +0000 UTC m=+794.027692902 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-memberlist") pod "speaker-hxwz6" (UID: "e316690b-6982-4588-ab38-a33ca86ba1f0") : secret "metallb-memberlist" not found Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.427839 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.436900 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-metrics-certs\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.438223 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf8mg\" (UniqueName: \"kubernetes.io/projected/e316690b-6982-4588-ab38-a33ca86ba1f0-kube-api-access-hf8mg\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.439085 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2473db3-776f-4a3d-9b4f-aba30d96463b-cert\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.439740 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2473db3-776f-4a3d-9b4f-aba30d96463b-metrics-certs\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.441921 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2pcb\" (UniqueName: \"kubernetes.io/projected/e2473db3-776f-4a3d-9b4f-aba30d96463b-kube-api-access-k2pcb\") pod \"controller-f8648f98b-6m7rl\" (UID: \"e2473db3-776f-4a3d-9b4f-aba30d96463b\") " pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.528415 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.587664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerStarted","Data":"1085d18ef7762c5da6b1e54f08c28d161e0bf01c25570c071f71383d06162a21"} Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.716221 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6m7rl"] Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.829029 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29388ab6-9e49-4662-8d74-1dda6e8d228e-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rt5h4\" (UID: \"29388ab6-9e49-4662-8d74-1dda6e8d228e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.833621 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29388ab6-9e49-4662-8d74-1dda6e8d228e-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-rt5h4\" (UID: \"29388ab6-9e49-4662-8d74-1dda6e8d228e\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.930140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-memberlist\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:32 crc kubenswrapper[4894]: I1208 14:59:32.934313 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e316690b-6982-4588-ab38-a33ca86ba1f0-memberlist\") pod \"speaker-hxwz6\" (UID: \"e316690b-6982-4588-ab38-a33ca86ba1f0\") " pod="metallb-system/speaker-hxwz6" Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.009838 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.069830 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hxwz6" Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.261282 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4"] Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.595268 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6m7rl" event={"ID":"e2473db3-776f-4a3d-9b4f-aba30d96463b","Type":"ContainerStarted","Data":"b6f7718839115c032573ba7cb30a0a8675afd26cb60bf023b710f13b2f29ecd5"} Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.595569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6m7rl" event={"ID":"e2473db3-776f-4a3d-9b4f-aba30d96463b","Type":"ContainerStarted","Data":"25c4fdaeadb582efdb9e5b5dee9c3f7a6b0542d3aa4332e98709d6428fd15cc3"} Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.595589 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.595598 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6m7rl" event={"ID":"e2473db3-776f-4a3d-9b4f-aba30d96463b","Type":"ContainerStarted","Data":"2dc762e90f79782b4b4af2d8758e5a1fafd04233fec014cad538920f799c8e60"} Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.596264 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hxwz6" event={"ID":"e316690b-6982-4588-ab38-a33ca86ba1f0","Type":"ContainerStarted","Data":"278b3d2acc1d5d07f823ca8ddd1f2fe4b499affd226b5682f03f519ec8020189"} Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.596302 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hxwz6" event={"ID":"e316690b-6982-4588-ab38-a33ca86ba1f0","Type":"ContainerStarted","Data":"ba34abc384133ca9fde397426f0ca0ad2bab5c5ad05df9f514564792386527f1"} Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.597256 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" event={"ID":"29388ab6-9e49-4662-8d74-1dda6e8d228e","Type":"ContainerStarted","Data":"3308003d4983badd34272d0e77d11ba24d63e126f370ff0fdde62704c13aabe1"} Dec 08 14:59:33 crc kubenswrapper[4894]: I1208 14:59:33.614304 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-6m7rl" podStartSLOduration=1.6142828150000001 podStartE2EDuration="1.614282815s" podCreationTimestamp="2025-12-08 14:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:59:33.61029732 +0000 UTC m=+794.710303445" watchObservedRunningTime="2025-12-08 14:59:33.614282815 +0000 UTC m=+794.714288930" Dec 08 14:59:34 crc kubenswrapper[4894]: I1208 14:59:34.620153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hxwz6" event={"ID":"e316690b-6982-4588-ab38-a33ca86ba1f0","Type":"ContainerStarted","Data":"917be1da665c4362354fc48e41d79e55152532d0c794f826a8a37ff60ba4de66"} Dec 08 14:59:34 crc kubenswrapper[4894]: I1208 14:59:34.620492 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hxwz6" Dec 08 14:59:39 crc kubenswrapper[4894]: I1208 14:59:39.226067 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hxwz6" podStartSLOduration=7.226046593 podStartE2EDuration="7.226046593s" podCreationTimestamp="2025-12-08 14:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 14:59:34.650211523 +0000 UTC m=+795.750217648" watchObservedRunningTime="2025-12-08 14:59:39.226046593 +0000 UTC m=+800.326052708" Dec 08 14:59:40 crc kubenswrapper[4894]: I1208 14:59:40.694571 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" event={"ID":"29388ab6-9e49-4662-8d74-1dda6e8d228e","Type":"ContainerStarted","Data":"d85b0886287a3eadd9db1025254dfc4582f71e5495deda25c88b0957305df690"} Dec 08 14:59:40 crc kubenswrapper[4894]: I1208 14:59:40.695166 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:40 crc kubenswrapper[4894]: I1208 14:59:40.696007 4894 generic.go:334] "Generic (PLEG): container finished" podID="e7389153-ceef-48e4-a814-b6204ce84627" containerID="53d36b4d1bca57b3341c9aecade5e48066deda05aeff69d40ebb6e8b59351dd4" exitCode=0 Dec 08 14:59:40 crc kubenswrapper[4894]: I1208 14:59:40.696050 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerDied","Data":"53d36b4d1bca57b3341c9aecade5e48066deda05aeff69d40ebb6e8b59351dd4"} Dec 08 14:59:40 crc kubenswrapper[4894]: I1208 14:59:40.710024 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" podStartSLOduration=1.790731742 podStartE2EDuration="8.710006655s" podCreationTimestamp="2025-12-08 14:59:32 +0000 UTC" firstStartedPulling="2025-12-08 14:59:33.277857597 +0000 UTC m=+794.377863712" lastFinishedPulling="2025-12-08 14:59:40.19713251 +0000 UTC m=+801.297138625" observedRunningTime="2025-12-08 14:59:40.707140136 +0000 UTC m=+801.807146241" watchObservedRunningTime="2025-12-08 14:59:40.710006655 +0000 UTC m=+801.810012770" Dec 08 14:59:41 crc kubenswrapper[4894]: I1208 14:59:41.701614 4894 generic.go:334] "Generic (PLEG): container finished" podID="e7389153-ceef-48e4-a814-b6204ce84627" containerID="2dba25f1870f19f0b4f7055ae8f1888508497620954426667569d47fc33823eb" exitCode=0 Dec 08 14:59:41 crc kubenswrapper[4894]: I1208 14:59:41.702780 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerDied","Data":"2dba25f1870f19f0b4f7055ae8f1888508497620954426667569d47fc33823eb"} Dec 08 14:59:42 crc kubenswrapper[4894]: I1208 14:59:42.534614 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-6m7rl" Dec 08 14:59:42 crc kubenswrapper[4894]: I1208 14:59:42.708270 4894 generic.go:334] "Generic (PLEG): container finished" podID="e7389153-ceef-48e4-a814-b6204ce84627" containerID="be8804139a44626521ab023633b625f546e81a9c5042576f2651d56862dbccb0" exitCode=0 Dec 08 14:59:42 crc kubenswrapper[4894]: I1208 14:59:42.708310 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerDied","Data":"be8804139a44626521ab023633b625f546e81a9c5042576f2651d56862dbccb0"} Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.074373 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hxwz6" Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.730805 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerStarted","Data":"496d95d71542880f60cee6028ff59e30c8f98afebb0c66feb93e7faf7961de0f"} Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.730880 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerStarted","Data":"03d3d095ba91eb6bd2af523d86dae3f92abb1bd6a9981820d0bdb292e8986043"} Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.730897 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerStarted","Data":"f4e2b13f39333a707afe43fbb97b272bc9682d7446471d26a44e911b7a4b1ab8"} Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.730905 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerStarted","Data":"dc177fcdb214a9b250edc47684296104ac65c2ccd6917ffaea5287437fa39031"} Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.730914 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerStarted","Data":"70a2178a7ccba3c996b426e9e2db93c9c1ce9a7be9e570adae0c92e32c21390c"} Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.730922 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5r5t2" event={"ID":"e7389153-ceef-48e4-a814-b6204ce84627","Type":"ContainerStarted","Data":"1c8ae30935569e9a0229f1f96608bb2a733bc0a48f28dfa6868ea82d0366c890"} Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.731095 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:43 crc kubenswrapper[4894]: I1208 14:59:43.754773 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-5r5t2" podStartSLOduration=4.121134016 podStartE2EDuration="11.754752116s" podCreationTimestamp="2025-12-08 14:59:32 +0000 UTC" firstStartedPulling="2025-12-08 14:59:32.544457684 +0000 UTC m=+793.644463799" lastFinishedPulling="2025-12-08 14:59:40.178075784 +0000 UTC m=+801.278081899" observedRunningTime="2025-12-08 14:59:43.751551245 +0000 UTC m=+804.851557360" watchObservedRunningTime="2025-12-08 14:59:43.754752116 +0000 UTC m=+804.854758231" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.055184 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-p78cg"] Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.056541 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p78cg" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.059340 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.059409 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-h5bw4" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.059541 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.070985 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-p78cg"] Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.204375 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kq7j\" (UniqueName: \"kubernetes.io/projected/adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58-kube-api-access-7kq7j\") pod \"openstack-operator-index-p78cg\" (UID: \"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58\") " pod="openstack-operators/openstack-operator-index-p78cg" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.305845 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kq7j\" (UniqueName: \"kubernetes.io/projected/adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58-kube-api-access-7kq7j\") pod \"openstack-operator-index-p78cg\" (UID: \"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58\") " pod="openstack-operators/openstack-operator-index-p78cg" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.324983 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kq7j\" (UniqueName: \"kubernetes.io/projected/adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58-kube-api-access-7kq7j\") pod \"openstack-operator-index-p78cg\" (UID: \"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58\") " pod="openstack-operators/openstack-operator-index-p78cg" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.375559 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p78cg" Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.576088 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-p78cg"] Dec 08 14:59:46 crc kubenswrapper[4894]: I1208 14:59:46.747534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p78cg" event={"ID":"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58","Type":"ContainerStarted","Data":"0b42945d71a92a48f49c715657615d765e319f5bf6e65a5a0e22917d6ac038b7"} Dec 08 14:59:47 crc kubenswrapper[4894]: I1208 14:59:47.401617 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:47 crc kubenswrapper[4894]: I1208 14:59:47.439177 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:49 crc kubenswrapper[4894]: I1208 14:59:49.434875 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-p78cg"] Dec 08 14:59:49 crc kubenswrapper[4894]: I1208 14:59:49.763610 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p78cg" event={"ID":"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58","Type":"ContainerStarted","Data":"556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a"} Dec 08 14:59:49 crc kubenswrapper[4894]: I1208 14:59:49.779752 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-p78cg" podStartSLOduration=1.205483754 podStartE2EDuration="3.779731267s" podCreationTimestamp="2025-12-08 14:59:46 +0000 UTC" firstStartedPulling="2025-12-08 14:59:46.584759981 +0000 UTC m=+807.684766096" lastFinishedPulling="2025-12-08 14:59:49.159007494 +0000 UTC m=+810.259013609" observedRunningTime="2025-12-08 14:59:49.778807547 +0000 UTC m=+810.878813682" watchObservedRunningTime="2025-12-08 14:59:49.779731267 +0000 UTC m=+810.879737382" Dec 08 14:59:49 crc kubenswrapper[4894]: I1208 14:59:49.843242 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-8ptfg"] Dec 08 14:59:49 crc kubenswrapper[4894]: I1208 14:59:49.844108 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 14:59:49 crc kubenswrapper[4894]: I1208 14:59:49.852022 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8ptfg"] Dec 08 14:59:49 crc kubenswrapper[4894]: I1208 14:59:49.951011 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt4bv\" (UniqueName: \"kubernetes.io/projected/654a1e08-5902-4a0f-8584-f662bc005210-kube-api-access-mt4bv\") pod \"openstack-operator-index-8ptfg\" (UID: \"654a1e08-5902-4a0f-8584-f662bc005210\") " pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.051943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt4bv\" (UniqueName: \"kubernetes.io/projected/654a1e08-5902-4a0f-8584-f662bc005210-kube-api-access-mt4bv\") pod \"openstack-operator-index-8ptfg\" (UID: \"654a1e08-5902-4a0f-8584-f662bc005210\") " pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.071028 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt4bv\" (UniqueName: \"kubernetes.io/projected/654a1e08-5902-4a0f-8584-f662bc005210-kube-api-access-mt4bv\") pod \"openstack-operator-index-8ptfg\" (UID: \"654a1e08-5902-4a0f-8584-f662bc005210\") " pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.157558 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.547439 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8ptfg"] Dec 08 14:59:50 crc kubenswrapper[4894]: W1208 14:59:50.552487 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod654a1e08_5902_4a0f_8584_f662bc005210.slice/crio-d7aeecfdf2d7e77ab3e7e0f6691267763a0fae1cddff55b685040d769d39f8cf WatchSource:0}: Error finding container d7aeecfdf2d7e77ab3e7e0f6691267763a0fae1cddff55b685040d769d39f8cf: Status 404 returned error can't find the container with id d7aeecfdf2d7e77ab3e7e0f6691267763a0fae1cddff55b685040d769d39f8cf Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.780419 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8ptfg" event={"ID":"654a1e08-5902-4a0f-8584-f662bc005210","Type":"ContainerStarted","Data":"9ddfa4951f4f40415482cc988f65fa753078e97b0a7a5a645c9cda3cd7d6af6a"} Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.781529 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8ptfg" event={"ID":"654a1e08-5902-4a0f-8584-f662bc005210","Type":"ContainerStarted","Data":"d7aeecfdf2d7e77ab3e7e0f6691267763a0fae1cddff55b685040d769d39f8cf"} Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.782323 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-p78cg" podUID="adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58" containerName="registry-server" containerID="cri-o://556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a" gracePeriod=2 Dec 08 14:59:50 crc kubenswrapper[4894]: I1208 14:59:50.806122 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-8ptfg" podStartSLOduration=1.76087776 podStartE2EDuration="1.806106006s" podCreationTimestamp="2025-12-08 14:59:49 +0000 UTC" firstStartedPulling="2025-12-08 14:59:50.554620859 +0000 UTC m=+811.654627004" lastFinishedPulling="2025-12-08 14:59:50.599849115 +0000 UTC m=+811.699855250" observedRunningTime="2025-12-08 14:59:50.800700376 +0000 UTC m=+811.900706511" watchObservedRunningTime="2025-12-08 14:59:50.806106006 +0000 UTC m=+811.906112121" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.271268 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p78cg" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.370675 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kq7j\" (UniqueName: \"kubernetes.io/projected/adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58-kube-api-access-7kq7j\") pod \"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58\" (UID: \"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58\") " Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.377763 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58-kube-api-access-7kq7j" (OuterVolumeSpecName: "kube-api-access-7kq7j") pod "adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58" (UID: "adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58"). InnerVolumeSpecName "kube-api-access-7kq7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.471850 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kq7j\" (UniqueName: \"kubernetes.io/projected/adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58-kube-api-access-7kq7j\") on node \"crc\" DevicePath \"\"" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.787280 4894 generic.go:334] "Generic (PLEG): container finished" podID="adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58" containerID="556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a" exitCode=0 Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.787347 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p78cg" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.787337 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p78cg" event={"ID":"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58","Type":"ContainerDied","Data":"556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a"} Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.787782 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p78cg" event={"ID":"adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58","Type":"ContainerDied","Data":"0b42945d71a92a48f49c715657615d765e319f5bf6e65a5a0e22917d6ac038b7"} Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.787807 4894 scope.go:117] "RemoveContainer" containerID="556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.803909 4894 scope.go:117] "RemoveContainer" containerID="556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a" Dec 08 14:59:51 crc kubenswrapper[4894]: E1208 14:59:51.804305 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a\": container with ID starting with 556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a not found: ID does not exist" containerID="556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.804348 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a"} err="failed to get container status \"556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a\": rpc error: code = NotFound desc = could not find container \"556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a\": container with ID starting with 556a5916e03679950f3e2de5e46b757134f6d22bd3f7e4a03881aeb1f8f4353a not found: ID does not exist" Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.816651 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-p78cg"] Dec 08 14:59:51 crc kubenswrapper[4894]: I1208 14:59:51.821609 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-p78cg"] Dec 08 14:59:52 crc kubenswrapper[4894]: I1208 14:59:52.403573 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-5r5t2" Dec 08 14:59:53 crc kubenswrapper[4894]: I1208 14:59:53.016449 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-rt5h4" Dec 08 14:59:53 crc kubenswrapper[4894]: I1208 14:59:53.205469 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58" path="/var/lib/kubelet/pods/adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58/volumes" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.153406 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk"] Dec 08 15:00:00 crc kubenswrapper[4894]: E1208 15:00:00.155283 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58" containerName="registry-server" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.155380 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58" containerName="registry-server" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.155572 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="adcd2558-2d52-4f74-b1b6-ed4fa1a2fb58" containerName="registry-server" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.156272 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.158549 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.158714 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.159172 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.162802 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.165398 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk"] Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.195539 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.278928 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7vl2\" (UniqueName: \"kubernetes.io/projected/b9f324fd-5585-43ac-9045-cc94aeb37fac-kube-api-access-b7vl2\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.278987 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9f324fd-5585-43ac-9045-cc94aeb37fac-secret-volume\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.279027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9f324fd-5585-43ac-9045-cc94aeb37fac-config-volume\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.380967 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7vl2\" (UniqueName: \"kubernetes.io/projected/b9f324fd-5585-43ac-9045-cc94aeb37fac-kube-api-access-b7vl2\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.381357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9f324fd-5585-43ac-9045-cc94aeb37fac-secret-volume\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.381404 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9f324fd-5585-43ac-9045-cc94aeb37fac-config-volume\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.382188 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9f324fd-5585-43ac-9045-cc94aeb37fac-config-volume\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.386505 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9f324fd-5585-43ac-9045-cc94aeb37fac-secret-volume\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.395220 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7vl2\" (UniqueName: \"kubernetes.io/projected/b9f324fd-5585-43ac-9045-cc94aeb37fac-kube-api-access-b7vl2\") pod \"collect-profiles-29420100-pgpxk\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.474332 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.901757 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-8ptfg" Dec 08 15:00:00 crc kubenswrapper[4894]: I1208 15:00:00.912218 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk"] Dec 08 15:00:01 crc kubenswrapper[4894]: I1208 15:00:01.847340 4894 generic.go:334] "Generic (PLEG): container finished" podID="b9f324fd-5585-43ac-9045-cc94aeb37fac" containerID="966763b2cd33c4d7db822327b65c6783eeed176e05227011693e39bf8f6abcfa" exitCode=0 Dec 08 15:00:01 crc kubenswrapper[4894]: I1208 15:00:01.847419 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" event={"ID":"b9f324fd-5585-43ac-9045-cc94aeb37fac","Type":"ContainerDied","Data":"966763b2cd33c4d7db822327b65c6783eeed176e05227011693e39bf8f6abcfa"} Dec 08 15:00:01 crc kubenswrapper[4894]: I1208 15:00:01.847718 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" event={"ID":"b9f324fd-5585-43ac-9045-cc94aeb37fac","Type":"ContainerStarted","Data":"1013e78bfb9b71b36cd914f84eb123ec1a5939bb7e44ce5e8c85d72ed607f31c"} Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.348020 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.517790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9f324fd-5585-43ac-9045-cc94aeb37fac-secret-volume\") pod \"b9f324fd-5585-43ac-9045-cc94aeb37fac\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.517902 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9f324fd-5585-43ac-9045-cc94aeb37fac-config-volume\") pod \"b9f324fd-5585-43ac-9045-cc94aeb37fac\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.517941 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7vl2\" (UniqueName: \"kubernetes.io/projected/b9f324fd-5585-43ac-9045-cc94aeb37fac-kube-api-access-b7vl2\") pod \"b9f324fd-5585-43ac-9045-cc94aeb37fac\" (UID: \"b9f324fd-5585-43ac-9045-cc94aeb37fac\") " Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.518733 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9f324fd-5585-43ac-9045-cc94aeb37fac-config-volume" (OuterVolumeSpecName: "config-volume") pod "b9f324fd-5585-43ac-9045-cc94aeb37fac" (UID: "b9f324fd-5585-43ac-9045-cc94aeb37fac"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.522872 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9f324fd-5585-43ac-9045-cc94aeb37fac-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b9f324fd-5585-43ac-9045-cc94aeb37fac" (UID: "b9f324fd-5585-43ac-9045-cc94aeb37fac"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.523096 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f324fd-5585-43ac-9045-cc94aeb37fac-kube-api-access-b7vl2" (OuterVolumeSpecName: "kube-api-access-b7vl2") pod "b9f324fd-5585-43ac-9045-cc94aeb37fac" (UID: "b9f324fd-5585-43ac-9045-cc94aeb37fac"). InnerVolumeSpecName "kube-api-access-b7vl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.619131 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9f324fd-5585-43ac-9045-cc94aeb37fac-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.619167 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9f324fd-5585-43ac-9045-cc94aeb37fac-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.619177 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7vl2\" (UniqueName: \"kubernetes.io/projected/b9f324fd-5585-43ac-9045-cc94aeb37fac-kube-api-access-b7vl2\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.862414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" event={"ID":"b9f324fd-5585-43ac-9045-cc94aeb37fac","Type":"ContainerDied","Data":"1013e78bfb9b71b36cd914f84eb123ec1a5939bb7e44ce5e8c85d72ed607f31c"} Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.862471 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1013e78bfb9b71b36cd914f84eb123ec1a5939bb7e44ce5e8c85d72ed607f31c" Dec 08 15:00:03 crc kubenswrapper[4894]: I1208 15:00:03.862495 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.672846 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk"] Dec 08 15:00:08 crc kubenswrapper[4894]: E1208 15:00:08.673371 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f324fd-5585-43ac-9045-cc94aeb37fac" containerName="collect-profiles" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.673384 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f324fd-5585-43ac-9045-cc94aeb37fac" containerName="collect-profiles" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.673503 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f324fd-5585-43ac-9045-cc94aeb37fac" containerName="collect-profiles" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.674386 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.676918 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-k94rj" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.689038 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk"] Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.781514 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-util\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.781577 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk82g\" (UniqueName: \"kubernetes.io/projected/11e95246-674f-4e1a-bfce-85a1e76658a1-kube-api-access-pk82g\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.781603 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-bundle\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.882806 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-bundle\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.883371 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-util\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.883605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk82g\" (UniqueName: \"kubernetes.io/projected/11e95246-674f-4e1a-bfce-85a1e76658a1-kube-api-access-pk82g\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.883701 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-bundle\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.883740 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-util\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.902217 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk82g\" (UniqueName: \"kubernetes.io/projected/11e95246-674f-4e1a-bfce-85a1e76658a1-kube-api-access-pk82g\") pod \"f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:08 crc kubenswrapper[4894]: I1208 15:00:08.990795 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:09 crc kubenswrapper[4894]: I1208 15:00:09.179930 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk"] Dec 08 15:00:09 crc kubenswrapper[4894]: W1208 15:00:09.181137 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11e95246_674f_4e1a_bfce_85a1e76658a1.slice/crio-7fc2ea2ea39dcf170698b906a89a3d3672a825db38c0c87819e68fa25eb9ad72 WatchSource:0}: Error finding container 7fc2ea2ea39dcf170698b906a89a3d3672a825db38c0c87819e68fa25eb9ad72: Status 404 returned error can't find the container with id 7fc2ea2ea39dcf170698b906a89a3d3672a825db38c0c87819e68fa25eb9ad72 Dec 08 15:00:09 crc kubenswrapper[4894]: I1208 15:00:09.900082 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" event={"ID":"11e95246-674f-4e1a-bfce-85a1e76658a1","Type":"ContainerStarted","Data":"7fc2ea2ea39dcf170698b906a89a3d3672a825db38c0c87819e68fa25eb9ad72"} Dec 08 15:00:11 crc kubenswrapper[4894]: I1208 15:00:11.918061 4894 generic.go:334] "Generic (PLEG): container finished" podID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerID="df0afe0ea0f741e57c65409cdf7f3003cc72f979c78e01a4cb0986b5eab5ec26" exitCode=0 Dec 08 15:00:11 crc kubenswrapper[4894]: I1208 15:00:11.918185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" event={"ID":"11e95246-674f-4e1a-bfce-85a1e76658a1","Type":"ContainerDied","Data":"df0afe0ea0f741e57c65409cdf7f3003cc72f979c78e01a4cb0986b5eab5ec26"} Dec 08 15:00:12 crc kubenswrapper[4894]: I1208 15:00:12.928171 4894 generic.go:334] "Generic (PLEG): container finished" podID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerID="3997749d4b2c1e5e52cf2b6259d65e8e0623b7c968b0b3593cfae2b7504b595a" exitCode=0 Dec 08 15:00:12 crc kubenswrapper[4894]: I1208 15:00:12.928258 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" event={"ID":"11e95246-674f-4e1a-bfce-85a1e76658a1","Type":"ContainerDied","Data":"3997749d4b2c1e5e52cf2b6259d65e8e0623b7c968b0b3593cfae2b7504b595a"} Dec 08 15:00:13 crc kubenswrapper[4894]: I1208 15:00:13.936285 4894 generic.go:334] "Generic (PLEG): container finished" podID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerID="721935d6cba6662bc1adf9367cb80b4b0fd0c1edd45f5d03ff53b620acce55d5" exitCode=0 Dec 08 15:00:13 crc kubenswrapper[4894]: I1208 15:00:13.936326 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" event={"ID":"11e95246-674f-4e1a-bfce-85a1e76658a1","Type":"ContainerDied","Data":"721935d6cba6662bc1adf9367cb80b4b0fd0c1edd45f5d03ff53b620acce55d5"} Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.234184 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.365286 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-bundle\") pod \"11e95246-674f-4e1a-bfce-85a1e76658a1\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.365397 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-util\") pod \"11e95246-674f-4e1a-bfce-85a1e76658a1\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.365440 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk82g\" (UniqueName: \"kubernetes.io/projected/11e95246-674f-4e1a-bfce-85a1e76658a1-kube-api-access-pk82g\") pod \"11e95246-674f-4e1a-bfce-85a1e76658a1\" (UID: \"11e95246-674f-4e1a-bfce-85a1e76658a1\") " Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.365998 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-bundle" (OuterVolumeSpecName: "bundle") pod "11e95246-674f-4e1a-bfce-85a1e76658a1" (UID: "11e95246-674f-4e1a-bfce-85a1e76658a1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.371801 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11e95246-674f-4e1a-bfce-85a1e76658a1-kube-api-access-pk82g" (OuterVolumeSpecName: "kube-api-access-pk82g") pod "11e95246-674f-4e1a-bfce-85a1e76658a1" (UID: "11e95246-674f-4e1a-bfce-85a1e76658a1"). InnerVolumeSpecName "kube-api-access-pk82g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.379677 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-util" (OuterVolumeSpecName: "util") pod "11e95246-674f-4e1a-bfce-85a1e76658a1" (UID: "11e95246-674f-4e1a-bfce-85a1e76658a1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.466728 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.466763 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e95246-674f-4e1a-bfce-85a1e76658a1-util\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.466775 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk82g\" (UniqueName: \"kubernetes.io/projected/11e95246-674f-4e1a-bfce-85a1e76658a1-kube-api-access-pk82g\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.950518 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" event={"ID":"11e95246-674f-4e1a-bfce-85a1e76658a1","Type":"ContainerDied","Data":"7fc2ea2ea39dcf170698b906a89a3d3672a825db38c0c87819e68fa25eb9ad72"} Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.950565 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fc2ea2ea39dcf170698b906a89a3d3672a825db38c0c87819e68fa25eb9ad72" Dec 08 15:00:15 crc kubenswrapper[4894]: I1208 15:00:15.950576 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.804616 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g"] Dec 08 15:00:20 crc kubenswrapper[4894]: E1208 15:00:20.805519 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerName="util" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.805538 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerName="util" Dec 08 15:00:20 crc kubenswrapper[4894]: E1208 15:00:20.805552 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerName="extract" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.805558 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerName="extract" Dec 08 15:00:20 crc kubenswrapper[4894]: E1208 15:00:20.805598 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerName="pull" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.805607 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerName="pull" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.805867 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="11e95246-674f-4e1a-bfce-85a1e76658a1" containerName="extract" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.806361 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.811175 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qs4rc" Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.833574 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g"] Dec 08 15:00:20 crc kubenswrapper[4894]: I1208 15:00:20.932257 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj2tb\" (UniqueName: \"kubernetes.io/projected/d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd-kube-api-access-zj2tb\") pod \"openstack-operator-controller-operator-647974bbd-56w9g\" (UID: \"d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd\") " pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" Dec 08 15:00:21 crc kubenswrapper[4894]: I1208 15:00:21.033718 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj2tb\" (UniqueName: \"kubernetes.io/projected/d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd-kube-api-access-zj2tb\") pod \"openstack-operator-controller-operator-647974bbd-56w9g\" (UID: \"d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd\") " pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" Dec 08 15:00:21 crc kubenswrapper[4894]: I1208 15:00:21.053144 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj2tb\" (UniqueName: \"kubernetes.io/projected/d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd-kube-api-access-zj2tb\") pod \"openstack-operator-controller-operator-647974bbd-56w9g\" (UID: \"d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd\") " pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" Dec 08 15:00:21 crc kubenswrapper[4894]: I1208 15:00:21.129564 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" Dec 08 15:00:21 crc kubenswrapper[4894]: I1208 15:00:21.337083 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g"] Dec 08 15:00:21 crc kubenswrapper[4894]: W1208 15:00:21.340550 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6d47a7f_7fcb_4b93_bccd_d6a79fd0e3dd.slice/crio-46197a0e48e7cc0c4f6cafbe52ac5c3f1b447a8f5a3513793ff94ee5ff3b542b WatchSource:0}: Error finding container 46197a0e48e7cc0c4f6cafbe52ac5c3f1b447a8f5a3513793ff94ee5ff3b542b: Status 404 returned error can't find the container with id 46197a0e48e7cc0c4f6cafbe52ac5c3f1b447a8f5a3513793ff94ee5ff3b542b Dec 08 15:00:21 crc kubenswrapper[4894]: I1208 15:00:21.991145 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" event={"ID":"d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd","Type":"ContainerStarted","Data":"46197a0e48e7cc0c4f6cafbe52ac5c3f1b447a8f5a3513793ff94ee5ff3b542b"} Dec 08 15:00:26 crc kubenswrapper[4894]: I1208 15:00:26.024570 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" event={"ID":"d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd","Type":"ContainerStarted","Data":"6c829e69d35fc9938227f3152c91590561d5e9128dedfaa67ba0b64a032c4ce7"} Dec 08 15:00:26 crc kubenswrapper[4894]: I1208 15:00:26.025148 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" Dec 08 15:00:26 crc kubenswrapper[4894]: I1208 15:00:26.054881 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" podStartSLOduration=1.949864806 podStartE2EDuration="6.054857163s" podCreationTimestamp="2025-12-08 15:00:20 +0000 UTC" firstStartedPulling="2025-12-08 15:00:21.342278482 +0000 UTC m=+842.442284597" lastFinishedPulling="2025-12-08 15:00:25.447270849 +0000 UTC m=+846.547276954" observedRunningTime="2025-12-08 15:00:26.049259237 +0000 UTC m=+847.149265352" watchObservedRunningTime="2025-12-08 15:00:26.054857163 +0000 UTC m=+847.154863278" Dec 08 15:00:31 crc kubenswrapper[4894]: I1208 15:00:31.132357 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-647974bbd-56w9g" Dec 08 15:00:36 crc kubenswrapper[4894]: I1208 15:00:36.801400 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6jp4n"] Dec 08 15:00:36 crc kubenswrapper[4894]: I1208 15:00:36.803757 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:36 crc kubenswrapper[4894]: I1208 15:00:36.814166 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jp4n"] Dec 08 15:00:36 crc kubenswrapper[4894]: I1208 15:00:36.934644 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xm9b\" (UniqueName: \"kubernetes.io/projected/6b5053cd-530c-44da-ba03-03b9bd509ba9-kube-api-access-8xm9b\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:36 crc kubenswrapper[4894]: I1208 15:00:36.935661 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-catalog-content\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:36 crc kubenswrapper[4894]: I1208 15:00:36.935850 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-utilities\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.036969 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xm9b\" (UniqueName: \"kubernetes.io/projected/6b5053cd-530c-44da-ba03-03b9bd509ba9-kube-api-access-8xm9b\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.037052 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-catalog-content\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.037079 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-utilities\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.037626 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-utilities\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.037878 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-catalog-content\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.064345 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xm9b\" (UniqueName: \"kubernetes.io/projected/6b5053cd-530c-44da-ba03-03b9bd509ba9-kube-api-access-8xm9b\") pod \"community-operators-6jp4n\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.126302 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:37 crc kubenswrapper[4894]: I1208 15:00:37.475063 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jp4n"] Dec 08 15:00:38 crc kubenswrapper[4894]: I1208 15:00:38.096238 4894 generic.go:334] "Generic (PLEG): container finished" podID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerID="972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7" exitCode=0 Dec 08 15:00:38 crc kubenswrapper[4894]: I1208 15:00:38.096334 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jp4n" event={"ID":"6b5053cd-530c-44da-ba03-03b9bd509ba9","Type":"ContainerDied","Data":"972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7"} Dec 08 15:00:38 crc kubenswrapper[4894]: I1208 15:00:38.096550 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jp4n" event={"ID":"6b5053cd-530c-44da-ba03-03b9bd509ba9","Type":"ContainerStarted","Data":"318cf11ae1a2e39d831b15e4b5baa14874870b9d5913482ca98454cae8ae5b24"} Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.104760 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jp4n" event={"ID":"6b5053cd-530c-44da-ba03-03b9bd509ba9","Type":"ContainerStarted","Data":"cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4"} Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.585891 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d9hgh"] Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.587582 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.599445 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9hgh"] Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.677930 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9qbq\" (UniqueName: \"kubernetes.io/projected/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-kube-api-access-s9qbq\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.678017 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-catalog-content\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.678092 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-utilities\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.779743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-catalog-content\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.780106 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-utilities\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.780191 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9qbq\" (UniqueName: \"kubernetes.io/projected/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-kube-api-access-s9qbq\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.780330 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-catalog-content\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.780575 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-utilities\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.799650 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9qbq\" (UniqueName: \"kubernetes.io/projected/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-kube-api-access-s9qbq\") pod \"redhat-marketplace-d9hgh\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:39 crc kubenswrapper[4894]: I1208 15:00:39.905072 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:40 crc kubenswrapper[4894]: I1208 15:00:40.113249 4894 generic.go:334] "Generic (PLEG): container finished" podID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerID="cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4" exitCode=0 Dec 08 15:00:40 crc kubenswrapper[4894]: I1208 15:00:40.113526 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jp4n" event={"ID":"6b5053cd-530c-44da-ba03-03b9bd509ba9","Type":"ContainerDied","Data":"cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4"} Dec 08 15:00:40 crc kubenswrapper[4894]: I1208 15:00:40.156215 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9hgh"] Dec 08 15:00:41 crc kubenswrapper[4894]: I1208 15:00:41.122140 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jp4n" event={"ID":"6b5053cd-530c-44da-ba03-03b9bd509ba9","Type":"ContainerStarted","Data":"c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336"} Dec 08 15:00:41 crc kubenswrapper[4894]: I1208 15:00:41.123358 4894 generic.go:334] "Generic (PLEG): container finished" podID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerID="1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56" exitCode=0 Dec 08 15:00:41 crc kubenswrapper[4894]: I1208 15:00:41.123399 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9hgh" event={"ID":"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4","Type":"ContainerDied","Data":"1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56"} Dec 08 15:00:41 crc kubenswrapper[4894]: I1208 15:00:41.123420 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9hgh" event={"ID":"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4","Type":"ContainerStarted","Data":"d3c02636a3c53c224d49c31ac5fdc16d664b92af765a3e907cac3180ceaba29b"} Dec 08 15:00:41 crc kubenswrapper[4894]: I1208 15:00:41.147828 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6jp4n" podStartSLOduration=2.712542504 podStartE2EDuration="5.147789354s" podCreationTimestamp="2025-12-08 15:00:36 +0000 UTC" firstStartedPulling="2025-12-08 15:00:38.097854427 +0000 UTC m=+859.197860542" lastFinishedPulling="2025-12-08 15:00:40.533101277 +0000 UTC m=+861.633107392" observedRunningTime="2025-12-08 15:00:41.142509408 +0000 UTC m=+862.242515553" watchObservedRunningTime="2025-12-08 15:00:41.147789354 +0000 UTC m=+862.247795469" Dec 08 15:00:42 crc kubenswrapper[4894]: I1208 15:00:42.144409 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9hgh" event={"ID":"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4","Type":"ContainerStarted","Data":"413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3"} Dec 08 15:00:43 crc kubenswrapper[4894]: I1208 15:00:43.150611 4894 generic.go:334] "Generic (PLEG): container finished" podID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerID="413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3" exitCode=0 Dec 08 15:00:43 crc kubenswrapper[4894]: I1208 15:00:43.150664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9hgh" event={"ID":"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4","Type":"ContainerDied","Data":"413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3"} Dec 08 15:00:44 crc kubenswrapper[4894]: I1208 15:00:44.158781 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9hgh" event={"ID":"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4","Type":"ContainerStarted","Data":"b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770"} Dec 08 15:00:44 crc kubenswrapper[4894]: I1208 15:00:44.189293 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d9hgh" podStartSLOduration=2.753881411 podStartE2EDuration="5.189274125s" podCreationTimestamp="2025-12-08 15:00:39 +0000 UTC" firstStartedPulling="2025-12-08 15:00:41.124400098 +0000 UTC m=+862.224406213" lastFinishedPulling="2025-12-08 15:00:43.559792812 +0000 UTC m=+864.659798927" observedRunningTime="2025-12-08 15:00:44.182590375 +0000 UTC m=+865.282596490" watchObservedRunningTime="2025-12-08 15:00:44.189274125 +0000 UTC m=+865.289280240" Dec 08 15:00:45 crc kubenswrapper[4894]: I1208 15:00:45.986294 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bsc4r"] Dec 08 15:00:45 crc kubenswrapper[4894]: I1208 15:00:45.990722 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:45 crc kubenswrapper[4894]: I1208 15:00:45.998868 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bsc4r"] Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.064306 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-utilities\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.064378 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-catalog-content\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.064400 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmdzd\" (UniqueName: \"kubernetes.io/projected/9e62b31c-801b-43fb-949f-3fd11c91604b-kube-api-access-zmdzd\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.165213 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-utilities\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.165313 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-catalog-content\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.165332 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmdzd\" (UniqueName: \"kubernetes.io/projected/9e62b31c-801b-43fb-949f-3fd11c91604b-kube-api-access-zmdzd\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.165910 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-utilities\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.166122 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-catalog-content\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.186498 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmdzd\" (UniqueName: \"kubernetes.io/projected/9e62b31c-801b-43fb-949f-3fd11c91604b-kube-api-access-zmdzd\") pod \"certified-operators-bsc4r\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.307260 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:46 crc kubenswrapper[4894]: I1208 15:00:46.624862 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bsc4r"] Dec 08 15:00:47 crc kubenswrapper[4894]: I1208 15:00:47.127462 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:47 crc kubenswrapper[4894]: I1208 15:00:47.127809 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:47 crc kubenswrapper[4894]: I1208 15:00:47.175365 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsc4r" event={"ID":"9e62b31c-801b-43fb-949f-3fd11c91604b","Type":"ContainerStarted","Data":"5d6a3888a8230d2547611aab4bd483034da44a52c7eada1d13eb0f884a615ed4"} Dec 08 15:00:47 crc kubenswrapper[4894]: I1208 15:00:47.215836 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:47 crc kubenswrapper[4894]: I1208 15:00:47.293444 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.193115 4894 generic.go:334] "Generic (PLEG): container finished" podID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerID="ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478" exitCode=0 Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.193213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsc4r" event={"ID":"9e62b31c-801b-43fb-949f-3fd11c91604b","Type":"ContainerDied","Data":"ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478"} Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.579726 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jp4n"] Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.580006 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6jp4n" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="registry-server" containerID="cri-o://c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336" gracePeriod=2 Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.906373 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.906508 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.989217 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:49 crc kubenswrapper[4894]: I1208 15:00:49.999753 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.121906 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-utilities\") pod \"6b5053cd-530c-44da-ba03-03b9bd509ba9\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.122546 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-utilities" (OuterVolumeSpecName: "utilities") pod "6b5053cd-530c-44da-ba03-03b9bd509ba9" (UID: "6b5053cd-530c-44da-ba03-03b9bd509ba9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.122700 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xm9b\" (UniqueName: \"kubernetes.io/projected/6b5053cd-530c-44da-ba03-03b9bd509ba9-kube-api-access-8xm9b\") pod \"6b5053cd-530c-44da-ba03-03b9bd509ba9\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.123517 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-catalog-content\") pod \"6b5053cd-530c-44da-ba03-03b9bd509ba9\" (UID: \"6b5053cd-530c-44da-ba03-03b9bd509ba9\") " Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.123806 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.128800 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5053cd-530c-44da-ba03-03b9bd509ba9-kube-api-access-8xm9b" (OuterVolumeSpecName: "kube-api-access-8xm9b") pod "6b5053cd-530c-44da-ba03-03b9bd509ba9" (UID: "6b5053cd-530c-44da-ba03-03b9bd509ba9"). InnerVolumeSpecName "kube-api-access-8xm9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.175632 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b5053cd-530c-44da-ba03-03b9bd509ba9" (UID: "6b5053cd-530c-44da-ba03-03b9bd509ba9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.202809 4894 generic.go:334] "Generic (PLEG): container finished" podID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerID="c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336" exitCode=0 Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.202895 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jp4n" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.202905 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jp4n" event={"ID":"6b5053cd-530c-44da-ba03-03b9bd509ba9","Type":"ContainerDied","Data":"c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336"} Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.202938 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jp4n" event={"ID":"6b5053cd-530c-44da-ba03-03b9bd509ba9","Type":"ContainerDied","Data":"318cf11ae1a2e39d831b15e4b5baa14874870b9d5913482ca98454cae8ae5b24"} Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.202962 4894 scope.go:117] "RemoveContainer" containerID="c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.206638 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsc4r" event={"ID":"9e62b31c-801b-43fb-949f-3fd11c91604b","Type":"ContainerStarted","Data":"add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299"} Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.222595 4894 scope.go:117] "RemoveContainer" containerID="cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.224966 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xm9b\" (UniqueName: \"kubernetes.io/projected/6b5053cd-530c-44da-ba03-03b9bd509ba9-kube-api-access-8xm9b\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.224993 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5053cd-530c-44da-ba03-03b9bd509ba9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.258415 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.259207 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jp4n"] Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.261864 4894 scope.go:117] "RemoveContainer" containerID="972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.266991 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6jp4n"] Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.277888 4894 scope.go:117] "RemoveContainer" containerID="c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336" Dec 08 15:00:50 crc kubenswrapper[4894]: E1208 15:00:50.281675 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336\": container with ID starting with c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336 not found: ID does not exist" containerID="c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.281737 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336"} err="failed to get container status \"c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336\": rpc error: code = NotFound desc = could not find container \"c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336\": container with ID starting with c565ffa777c26da7767a30a500d1b1a03737db9a7c6c2229de5068126efd0336 not found: ID does not exist" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.281780 4894 scope.go:117] "RemoveContainer" containerID="cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4" Dec 08 15:00:50 crc kubenswrapper[4894]: E1208 15:00:50.282337 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4\": container with ID starting with cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4 not found: ID does not exist" containerID="cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.282373 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4"} err="failed to get container status \"cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4\": rpc error: code = NotFound desc = could not find container \"cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4\": container with ID starting with cab14ec9fec7601732cffc39b5865094e47bbcdb979cbcba305d18d80e3b88e4 not found: ID does not exist" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.282400 4894 scope.go:117] "RemoveContainer" containerID="972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7" Dec 08 15:00:50 crc kubenswrapper[4894]: E1208 15:00:50.283316 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7\": container with ID starting with 972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7 not found: ID does not exist" containerID="972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7" Dec 08 15:00:50 crc kubenswrapper[4894]: I1208 15:00:50.283353 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7"} err="failed to get container status \"972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7\": rpc error: code = NotFound desc = could not find container \"972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7\": container with ID starting with 972b213f01140ee77352f9395ffc946b7c415b15b710d37b9409d2e3c2e2bfe7 not found: ID does not exist" Dec 08 15:00:51 crc kubenswrapper[4894]: I1208 15:00:51.203722 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" path="/var/lib/kubelet/pods/6b5053cd-530c-44da-ba03-03b9bd509ba9/volumes" Dec 08 15:00:51 crc kubenswrapper[4894]: I1208 15:00:51.216993 4894 generic.go:334] "Generic (PLEG): container finished" podID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerID="add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299" exitCode=0 Dec 08 15:00:51 crc kubenswrapper[4894]: I1208 15:00:51.217026 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsc4r" event={"ID":"9e62b31c-801b-43fb-949f-3fd11c91604b","Type":"ContainerDied","Data":"add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299"} Dec 08 15:00:52 crc kubenswrapper[4894]: I1208 15:00:52.224583 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsc4r" event={"ID":"9e62b31c-801b-43fb-949f-3fd11c91604b","Type":"ContainerStarted","Data":"fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957"} Dec 08 15:00:52 crc kubenswrapper[4894]: I1208 15:00:52.242270 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bsc4r" podStartSLOduration=4.6516982460000005 podStartE2EDuration="7.242252369s" podCreationTimestamp="2025-12-08 15:00:45 +0000 UTC" firstStartedPulling="2025-12-08 15:00:49.195809219 +0000 UTC m=+870.295815334" lastFinishedPulling="2025-12-08 15:00:51.786363342 +0000 UTC m=+872.886369457" observedRunningTime="2025-12-08 15:00:52.237421097 +0000 UTC m=+873.337427222" watchObservedRunningTime="2025-12-08 15:00:52.242252369 +0000 UTC m=+873.342258484" Dec 08 15:00:53 crc kubenswrapper[4894]: I1208 15:00:53.777747 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9hgh"] Dec 08 15:00:53 crc kubenswrapper[4894]: I1208 15:00:53.778298 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d9hgh" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="registry-server" containerID="cri-o://b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770" gracePeriod=2 Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.119030 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.175232 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-catalog-content\") pod \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.175387 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9qbq\" (UniqueName: \"kubernetes.io/projected/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-kube-api-access-s9qbq\") pod \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.175424 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-utilities\") pod \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\" (UID: \"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4\") " Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.176486 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-utilities" (OuterVolumeSpecName: "utilities") pod "f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" (UID: "f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.192672 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-kube-api-access-s9qbq" (OuterVolumeSpecName: "kube-api-access-s9qbq") pod "f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" (UID: "f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4"). InnerVolumeSpecName "kube-api-access-s9qbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.211678 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" (UID: "f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.245751 4894 generic.go:334] "Generic (PLEG): container finished" podID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerID="b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770" exitCode=0 Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.245786 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d9hgh" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.245795 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9hgh" event={"ID":"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4","Type":"ContainerDied","Data":"b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770"} Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.245874 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d9hgh" event={"ID":"f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4","Type":"ContainerDied","Data":"d3c02636a3c53c224d49c31ac5fdc16d664b92af765a3e907cac3180ceaba29b"} Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.245917 4894 scope.go:117] "RemoveContainer" containerID="b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.266213 4894 scope.go:117] "RemoveContainer" containerID="413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.277798 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9qbq\" (UniqueName: \"kubernetes.io/projected/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-kube-api-access-s9qbq\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.277864 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.277880 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.277985 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9hgh"] Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.281699 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d9hgh"] Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.291630 4894 scope.go:117] "RemoveContainer" containerID="1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.313704 4894 scope.go:117] "RemoveContainer" containerID="b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770" Dec 08 15:00:54 crc kubenswrapper[4894]: E1208 15:00:54.314566 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770\": container with ID starting with b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770 not found: ID does not exist" containerID="b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.314613 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770"} err="failed to get container status \"b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770\": rpc error: code = NotFound desc = could not find container \"b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770\": container with ID starting with b61e65255db1bd6b55aa8254bc3934aaab848adbc946456d5f31a09c310cd770 not found: ID does not exist" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.314642 4894 scope.go:117] "RemoveContainer" containerID="413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3" Dec 08 15:00:54 crc kubenswrapper[4894]: E1208 15:00:54.315372 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3\": container with ID starting with 413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3 not found: ID does not exist" containerID="413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.315446 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3"} err="failed to get container status \"413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3\": rpc error: code = NotFound desc = could not find container \"413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3\": container with ID starting with 413996e3738e9731c56c027eda04b1c777dc7a2ed3137208818a8e740c0f72e3 not found: ID does not exist" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.315492 4894 scope.go:117] "RemoveContainer" containerID="1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56" Dec 08 15:00:54 crc kubenswrapper[4894]: E1208 15:00:54.315863 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56\": container with ID starting with 1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56 not found: ID does not exist" containerID="1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56" Dec 08 15:00:54 crc kubenswrapper[4894]: I1208 15:00:54.315894 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56"} err="failed to get container status \"1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56\": rpc error: code = NotFound desc = could not find container \"1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56\": container with ID starting with 1dd7e928ae8771b0b7f5285911c91d35b7c1523d6dcf6b83574b352db43f4b56 not found: ID does not exist" Dec 08 15:00:55 crc kubenswrapper[4894]: I1208 15:00:55.205796 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" path="/var/lib/kubelet/pods/f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4/volumes" Dec 08 15:00:56 crc kubenswrapper[4894]: I1208 15:00:56.307872 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:56 crc kubenswrapper[4894]: I1208 15:00:56.307927 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:56 crc kubenswrapper[4894]: I1208 15:00:56.399408 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:57 crc kubenswrapper[4894]: I1208 15:00:57.310165 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:57 crc kubenswrapper[4894]: I1208 15:00:57.978068 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bsc4r"] Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.271088 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bsc4r" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="registry-server" containerID="cri-o://fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957" gracePeriod=2 Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.612260 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.745728 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-catalog-content\") pod \"9e62b31c-801b-43fb-949f-3fd11c91604b\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.745830 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-utilities\") pod \"9e62b31c-801b-43fb-949f-3fd11c91604b\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.745878 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmdzd\" (UniqueName: \"kubernetes.io/projected/9e62b31c-801b-43fb-949f-3fd11c91604b-kube-api-access-zmdzd\") pod \"9e62b31c-801b-43fb-949f-3fd11c91604b\" (UID: \"9e62b31c-801b-43fb-949f-3fd11c91604b\") " Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.747087 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-utilities" (OuterVolumeSpecName: "utilities") pod "9e62b31c-801b-43fb-949f-3fd11c91604b" (UID: "9e62b31c-801b-43fb-949f-3fd11c91604b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.759969 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e62b31c-801b-43fb-949f-3fd11c91604b-kube-api-access-zmdzd" (OuterVolumeSpecName: "kube-api-access-zmdzd") pod "9e62b31c-801b-43fb-949f-3fd11c91604b" (UID: "9e62b31c-801b-43fb-949f-3fd11c91604b"). InnerVolumeSpecName "kube-api-access-zmdzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.814110 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e62b31c-801b-43fb-949f-3fd11c91604b" (UID: "9e62b31c-801b-43fb-949f-3fd11c91604b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.847529 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmdzd\" (UniqueName: \"kubernetes.io/projected/9e62b31c-801b-43fb-949f-3fd11c91604b-kube-api-access-zmdzd\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.847780 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:00:59 crc kubenswrapper[4894]: I1208 15:00:59.847903 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e62b31c-801b-43fb-949f-3fd11c91604b-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.278275 4894 generic.go:334] "Generic (PLEG): container finished" podID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerID="fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957" exitCode=0 Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.278314 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsc4r" event={"ID":"9e62b31c-801b-43fb-949f-3fd11c91604b","Type":"ContainerDied","Data":"fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957"} Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.278338 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bsc4r" event={"ID":"9e62b31c-801b-43fb-949f-3fd11c91604b","Type":"ContainerDied","Data":"5d6a3888a8230d2547611aab4bd483034da44a52c7eada1d13eb0f884a615ed4"} Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.278353 4894 scope.go:117] "RemoveContainer" containerID="fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.278479 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bsc4r" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.308653 4894 scope.go:117] "RemoveContainer" containerID="add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.312621 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bsc4r"] Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.316913 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bsc4r"] Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.340195 4894 scope.go:117] "RemoveContainer" containerID="ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.355620 4894 scope.go:117] "RemoveContainer" containerID="fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957" Dec 08 15:01:00 crc kubenswrapper[4894]: E1208 15:01:00.356045 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957\": container with ID starting with fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957 not found: ID does not exist" containerID="fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.356078 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957"} err="failed to get container status \"fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957\": rpc error: code = NotFound desc = could not find container \"fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957\": container with ID starting with fa6f12dd272943bca82b55785484d2649078726d5efb026c309b90e08c02c957 not found: ID does not exist" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.356108 4894 scope.go:117] "RemoveContainer" containerID="add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299" Dec 08 15:01:00 crc kubenswrapper[4894]: E1208 15:01:00.356433 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299\": container with ID starting with add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299 not found: ID does not exist" containerID="add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.356456 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299"} err="failed to get container status \"add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299\": rpc error: code = NotFound desc = could not find container \"add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299\": container with ID starting with add4b07c1a511b607625ac487414f5beda9e5763084267ed440997b92dd01299 not found: ID does not exist" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.356471 4894 scope.go:117] "RemoveContainer" containerID="ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478" Dec 08 15:01:00 crc kubenswrapper[4894]: E1208 15:01:00.356786 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478\": container with ID starting with ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478 not found: ID does not exist" containerID="ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478" Dec 08 15:01:00 crc kubenswrapper[4894]: I1208 15:01:00.356844 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478"} err="failed to get container status \"ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478\": rpc error: code = NotFound desc = could not find container \"ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478\": container with ID starting with ef725257c2131539da8dafbf77c26988147083e6e98f0f8088785849496d6478 not found: ID does not exist" Dec 08 15:01:01 crc kubenswrapper[4894]: I1208 15:01:01.205039 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" path="/var/lib/kubelet/pods/9e62b31c-801b-43fb-949f-3fd11c91604b/volumes" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.427513 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r6hn7"] Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428344 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="extract-utilities" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428362 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="extract-utilities" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428373 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="extract-content" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428381 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="extract-content" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428392 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="extract-utilities" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428400 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="extract-utilities" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428412 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="extract-utilities" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428421 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="extract-utilities" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428437 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="extract-content" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428446 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="extract-content" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428455 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="extract-content" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428463 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="extract-content" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428480 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428489 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428499 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428507 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: E1208 15:01:07.428520 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428528 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428659 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e62b31c-801b-43fb-949f-3fd11c91604b" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428681 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0a7ff4f-d93c-4eb4-a266-b29ba262c7f4" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.428690 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5053cd-530c-44da-ba03-03b9bd509ba9" containerName="registry-server" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.429758 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.440660 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r6hn7"] Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.544282 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-utilities\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.544358 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-catalog-content\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.544376 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfp94\" (UniqueName: \"kubernetes.io/projected/a377a91d-b683-4a28-9633-0c3f2367e3e5-kube-api-access-sfp94\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.645577 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-utilities\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.645656 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-catalog-content\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.645676 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfp94\" (UniqueName: \"kubernetes.io/projected/a377a91d-b683-4a28-9633-0c3f2367e3e5-kube-api-access-sfp94\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.646266 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-utilities\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.646272 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-catalog-content\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.671780 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfp94\" (UniqueName: \"kubernetes.io/projected/a377a91d-b683-4a28-9633-0c3f2367e3e5-kube-api-access-sfp94\") pod \"redhat-operators-r6hn7\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:07 crc kubenswrapper[4894]: I1208 15:01:07.749467 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.065363 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r6hn7"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.256533 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.257653 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.262339 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-9q4cp" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.269303 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.274742 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.275808 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.277464 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-kjnhr" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.289584 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.290804 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.293933 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-kwj4t" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.297892 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.298958 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.300547 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gzwmz" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.306705 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.310793 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.320006 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.320946 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.324621 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8cz2x" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.326038 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerStarted","Data":"cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d"} Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.326086 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerStarted","Data":"bdfbdfb6ce441d1ad4fdee3f2f6cd28b5268ade8781d82a50a00f1d75c7bc441"} Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.328990 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.355043 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84klp\" (UniqueName: \"kubernetes.io/projected/b8a908e6-3df2-4dc9-9c9c-068e91dc839c-kube-api-access-84klp\") pod \"glance-operator-controller-manager-5697bb5779-ffvpw\" (UID: \"b8a908e6-3df2-4dc9-9c9c-068e91dc839c\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.355429 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxqq4\" (UniqueName: \"kubernetes.io/projected/4ee4db87-8c97-4002-844b-6c55e1a41b2e-kube-api-access-gxqq4\") pod \"cinder-operator-controller-manager-6c677c69b-8vb7c\" (UID: \"4ee4db87-8c97-4002-844b-6c55e1a41b2e\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.355502 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xbq4\" (UniqueName: \"kubernetes.io/projected/28cfecb9-9c7c-49ae-af33-fc2d9acd374a-kube-api-access-2xbq4\") pod \"barbican-operator-controller-manager-7d9dfd778-kh5v9\" (UID: \"28cfecb9-9c7c-49ae-af33-fc2d9acd374a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.355550 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6jlp\" (UniqueName: \"kubernetes.io/projected/e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72-kube-api-access-z6jlp\") pod \"designate-operator-controller-manager-697fb699cf-f67zc\" (UID: \"e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.360979 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.367396 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.368462 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.379893 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.381114 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.382183 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.382239 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-964zj" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.383904 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-47bpp"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.385033 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.388495 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.391239 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-kg9b8" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.394657 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-42rzn" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.424599 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.439976 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-47bpp"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456504 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456596 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xbq4\" (UniqueName: \"kubernetes.io/projected/28cfecb9-9c7c-49ae-af33-fc2d9acd374a-kube-api-access-2xbq4\") pod \"barbican-operator-controller-manager-7d9dfd778-kh5v9\" (UID: \"28cfecb9-9c7c-49ae-af33-fc2d9acd374a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456632 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2bvd\" (UniqueName: \"kubernetes.io/projected/c3e89da4-2a46-4982-9392-b0cf87a1edc3-kube-api-access-n2bvd\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456664 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6jlp\" (UniqueName: \"kubernetes.io/projected/e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72-kube-api-access-z6jlp\") pod \"designate-operator-controller-manager-697fb699cf-f67zc\" (UID: \"e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456747 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84klp\" (UniqueName: \"kubernetes.io/projected/b8a908e6-3df2-4dc9-9c9c-068e91dc839c-kube-api-access-84klp\") pod \"glance-operator-controller-manager-5697bb5779-ffvpw\" (UID: \"b8a908e6-3df2-4dc9-9c9c-068e91dc839c\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456782 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw67p\" (UniqueName: \"kubernetes.io/projected/afda731d-73a8-443f-9647-a774ca8c3146-kube-api-access-dw67p\") pod \"horizon-operator-controller-manager-68c6d99b8f-mtkrb\" (UID: \"afda731d-73a8-443f-9647-a774ca8c3146\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456803 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpjrw\" (UniqueName: \"kubernetes.io/projected/5d7b9bb8-c977-45cb-b198-5f047059a0a4-kube-api-access-rpjrw\") pod \"heat-operator-controller-manager-5f64f6f8bb-frz8t\" (UID: \"5d7b9bb8-c977-45cb-b198-5f047059a0a4\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456847 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcphm\" (UniqueName: \"kubernetes.io/projected/f7ec7a43-f5b5-474f-af15-d0e664699477-kube-api-access-gcphm\") pod \"ironic-operator-controller-manager-967d97867-47bpp\" (UID: \"f7ec7a43-f5b5-474f-af15-d0e664699477\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.456871 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxqq4\" (UniqueName: \"kubernetes.io/projected/4ee4db87-8c97-4002-844b-6c55e1a41b2e-kube-api-access-gxqq4\") pod \"cinder-operator-controller-manager-6c677c69b-8vb7c\" (UID: \"4ee4db87-8c97-4002-844b-6c55e1a41b2e\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.481001 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.482234 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.484889 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-5zrn2" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.505695 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84klp\" (UniqueName: \"kubernetes.io/projected/b8a908e6-3df2-4dc9-9c9c-068e91dc839c-kube-api-access-84klp\") pod \"glance-operator-controller-manager-5697bb5779-ffvpw\" (UID: \"b8a908e6-3df2-4dc9-9c9c-068e91dc839c\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.506982 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xbq4\" (UniqueName: \"kubernetes.io/projected/28cfecb9-9c7c-49ae-af33-fc2d9acd374a-kube-api-access-2xbq4\") pod \"barbican-operator-controller-manager-7d9dfd778-kh5v9\" (UID: \"28cfecb9-9c7c-49ae-af33-fc2d9acd374a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.507631 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6jlp\" (UniqueName: \"kubernetes.io/projected/e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72-kube-api-access-z6jlp\") pod \"designate-operator-controller-manager-697fb699cf-f67zc\" (UID: \"e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.519197 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxqq4\" (UniqueName: \"kubernetes.io/projected/4ee4db87-8c97-4002-844b-6c55e1a41b2e-kube-api-access-gxqq4\") pod \"cinder-operator-controller-manager-6c677c69b-8vb7c\" (UID: \"4ee4db87-8c97-4002-844b-6c55e1a41b2e\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.560071 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6hp\" (UniqueName: \"kubernetes.io/projected/aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2-kube-api-access-7c6hp\") pod \"manila-operator-controller-manager-5b5fd79c9c-w7xbj\" (UID: \"aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.560186 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2bvd\" (UniqueName: \"kubernetes.io/projected/c3e89da4-2a46-4982-9392-b0cf87a1edc3-kube-api-access-n2bvd\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.560647 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw67p\" (UniqueName: \"kubernetes.io/projected/afda731d-73a8-443f-9647-a774ca8c3146-kube-api-access-dw67p\") pod \"horizon-operator-controller-manager-68c6d99b8f-mtkrb\" (UID: \"afda731d-73a8-443f-9647-a774ca8c3146\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.560677 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpjrw\" (UniqueName: \"kubernetes.io/projected/5d7b9bb8-c977-45cb-b198-5f047059a0a4-kube-api-access-rpjrw\") pod \"heat-operator-controller-manager-5f64f6f8bb-frz8t\" (UID: \"5d7b9bb8-c977-45cb-b198-5f047059a0a4\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.560702 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcphm\" (UniqueName: \"kubernetes.io/projected/f7ec7a43-f5b5-474f-af15-d0e664699477-kube-api-access-gcphm\") pod \"ironic-operator-controller-manager-967d97867-47bpp\" (UID: \"f7ec7a43-f5b5-474f-af15-d0e664699477\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.561582 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:08 crc kubenswrapper[4894]: E1208 15:01:08.561963 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:08 crc kubenswrapper[4894]: E1208 15:01:08.562031 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert podName:c3e89da4-2a46-4982-9392-b0cf87a1edc3 nodeName:}" failed. No retries permitted until 2025-12-08 15:01:09.062015806 +0000 UTC m=+890.162021921 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert") pod "infra-operator-controller-manager-78d48bff9d-4xrts" (UID: "c3e89da4-2a46-4982-9392-b0cf87a1edc3") : secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.578238 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.602755 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2bvd\" (UniqueName: \"kubernetes.io/projected/c3e89da4-2a46-4982-9392-b0cf87a1edc3-kube-api-access-n2bvd\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.603557 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcphm\" (UniqueName: \"kubernetes.io/projected/f7ec7a43-f5b5-474f-af15-d0e664699477-kube-api-access-gcphm\") pod \"ironic-operator-controller-manager-967d97867-47bpp\" (UID: \"f7ec7a43-f5b5-474f-af15-d0e664699477\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.654111 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpjrw\" (UniqueName: \"kubernetes.io/projected/5d7b9bb8-c977-45cb-b198-5f047059a0a4-kube-api-access-rpjrw\") pod \"heat-operator-controller-manager-5f64f6f8bb-frz8t\" (UID: \"5d7b9bb8-c977-45cb-b198-5f047059a0a4\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.659501 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw67p\" (UniqueName: \"kubernetes.io/projected/afda731d-73a8-443f-9647-a774ca8c3146-kube-api-access-dw67p\") pod \"horizon-operator-controller-manager-68c6d99b8f-mtkrb\" (UID: \"afda731d-73a8-443f-9647-a774ca8c3146\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.665791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6hp\" (UniqueName: \"kubernetes.io/projected/aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2-kube-api-access-7c6hp\") pod \"manila-operator-controller-manager-5b5fd79c9c-w7xbj\" (UID: \"aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.690047 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.691301 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.705303 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-smjjb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.705306 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.711532 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6hp\" (UniqueName: \"kubernetes.io/projected/aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2-kube-api-access-7c6hp\") pod \"manila-operator-controller-manager-5b5fd79c9c-w7xbj\" (UID: \"aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.719318 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.730589 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.733579 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.742510 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.752087 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.771904 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dtln\" (UniqueName: \"kubernetes.io/projected/2df93571-3b4e-4ab0-ad8b-34d2ef76034c-kube-api-access-9dtln\") pod \"keystone-operator-controller-manager-7765d96ddf-lbszt\" (UID: \"2df93571-3b4e-4ab0-ad8b-34d2ef76034c\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.784771 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.786011 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.797458 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-cqs2s" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.799992 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.801176 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.803026 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-54kdk" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.841703 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.859586 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.864032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.868950 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.875259 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-hqppc" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.876213 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm7md\" (UniqueName: \"kubernetes.io/projected/e6585a6c-b50d-41d3-a080-ab7cb27f9199-kube-api-access-bm7md\") pod \"mariadb-operator-controller-manager-79c8c4686c-h2xw4\" (UID: \"e6585a6c-b50d-41d3-a080-ab7cb27f9199\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.876279 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jvwl\" (UniqueName: \"kubernetes.io/projected/db804bba-5a84-4d23-93ca-e7118afc2af8-kube-api-access-8jvwl\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2sfqd\" (UID: \"db804bba-5a84-4d23-93ca-e7118afc2af8\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.876339 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dtln\" (UniqueName: \"kubernetes.io/projected/2df93571-3b4e-4ab0-ad8b-34d2ef76034c-kube-api-access-9dtln\") pod \"keystone-operator-controller-manager-7765d96ddf-lbszt\" (UID: \"2df93571-3b4e-4ab0-ad8b-34d2ef76034c\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.876418 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bvvz\" (UniqueName: \"kubernetes.io/projected/c91237dd-8bb9-4240-b24a-e19b58687d2e-kube-api-access-7bvvz\") pod \"nova-operator-controller-manager-697bc559fc-jc2rb\" (UID: \"c91237dd-8bb9-4240-b24a-e19b58687d2e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.903667 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.906077 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.943870 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.957043 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dtln\" (UniqueName: \"kubernetes.io/projected/2df93571-3b4e-4ab0-ad8b-34d2ef76034c-kube-api-access-9dtln\") pod \"keystone-operator-controller-manager-7765d96ddf-lbszt\" (UID: \"2df93571-3b4e-4ab0-ad8b-34d2ef76034c\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.955379 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.958220 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.961598 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-kljqw" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.966025 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb"] Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.978364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm7md\" (UniqueName: \"kubernetes.io/projected/e6585a6c-b50d-41d3-a080-ab7cb27f9199-kube-api-access-bm7md\") pod \"mariadb-operator-controller-manager-79c8c4686c-h2xw4\" (UID: \"e6585a6c-b50d-41d3-a080-ab7cb27f9199\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.978457 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jvwl\" (UniqueName: \"kubernetes.io/projected/db804bba-5a84-4d23-93ca-e7118afc2af8-kube-api-access-8jvwl\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2sfqd\" (UID: \"db804bba-5a84-4d23-93ca-e7118afc2af8\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.978573 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhfdk\" (UniqueName: \"kubernetes.io/projected/750c567a-693a-401b-8e7f-d8f1befbeacb-kube-api-access-zhfdk\") pod \"octavia-operator-controller-manager-998648c74-cqq4p\" (UID: \"750c567a-693a-401b-8e7f-d8f1befbeacb\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.978651 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bvvz\" (UniqueName: \"kubernetes.io/projected/c91237dd-8bb9-4240-b24a-e19b58687d2e-kube-api-access-7bvvz\") pod \"nova-operator-controller-manager-697bc559fc-jc2rb\" (UID: \"c91237dd-8bb9-4240-b24a-e19b58687d2e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" Dec 08 15:01:08 crc kubenswrapper[4894]: I1208 15:01:08.980213 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.022685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jvwl\" (UniqueName: \"kubernetes.io/projected/db804bba-5a84-4d23-93ca-e7118afc2af8-kube-api-access-8jvwl\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2sfqd\" (UID: \"db804bba-5a84-4d23-93ca-e7118afc2af8\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.021763 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.025376 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.029329 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5gjs4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.029518 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.030396 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm7md\" (UniqueName: \"kubernetes.io/projected/e6585a6c-b50d-41d3-a080-ab7cb27f9199-kube-api-access-bm7md\") pod \"mariadb-operator-controller-manager-79c8c4686c-h2xw4\" (UID: \"e6585a6c-b50d-41d3-a080-ab7cb27f9199\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.031389 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.034743 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dff46" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.037287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bvvz\" (UniqueName: \"kubernetes.io/projected/c91237dd-8bb9-4240-b24a-e19b58687d2e-kube-api-access-7bvvz\") pod \"nova-operator-controller-manager-697bc559fc-jc2rb\" (UID: \"c91237dd-8bb9-4240-b24a-e19b58687d2e\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.048059 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.059711 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.070110 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.076516 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.078922 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.079039 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.080332 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.080428 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhfdk\" (UniqueName: \"kubernetes.io/projected/750c567a-693a-401b-8e7f-d8f1befbeacb-kube-api-access-zhfdk\") pod \"octavia-operator-controller-manager-998648c74-cqq4p\" (UID: \"750c567a-693a-401b-8e7f-d8f1befbeacb\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.080780 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.080888 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert podName:c3e89da4-2a46-4982-9392-b0cf87a1edc3 nodeName:}" failed. No retries permitted until 2025-12-08 15:01:10.080812194 +0000 UTC m=+891.180818309 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert") pod "infra-operator-controller-manager-78d48bff9d-4xrts" (UID: "c3e89da4-2a46-4982-9392-b0cf87a1edc3") : secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.085659 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-l2cm7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.099692 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.100760 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.106586 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.108116 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-k2btj" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.108341 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.109670 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.112174 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhfdk\" (UniqueName: \"kubernetes.io/projected/750c567a-693a-401b-8e7f-d8f1befbeacb-kube-api-access-zhfdk\") pod \"octavia-operator-controller-manager-998648c74-cqq4p\" (UID: \"750c567a-693a-401b-8e7f-d8f1befbeacb\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.121157 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-klk8n" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.121746 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.142437 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.184509 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-279hf\" (UniqueName: \"kubernetes.io/projected/2b5f3682-eb12-440c-a016-f165ec4e7cfa-kube-api-access-279hf\") pod \"ovn-operator-controller-manager-b6456fdb6-lr526\" (UID: \"2b5f3682-eb12-440c-a016-f165ec4e7cfa\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.184632 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4gj5\" (UniqueName: \"kubernetes.io/projected/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-kube-api-access-g4gj5\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.184673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwsxz\" (UniqueName: \"kubernetes.io/projected/59051068-3252-47e6-a826-79d1951c3d5c-kube-api-access-jwsxz\") pod \"placement-operator-controller-manager-78f8948974-sxfkp\" (UID: \"59051068-3252-47e6-a826-79d1951c3d5c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.184715 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.199315 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.265992 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.299785 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw5b6\" (UniqueName: \"kubernetes.io/projected/2536d7d4-e60f-44a6-8314-cfd34b5545e9-kube-api-access-gw5b6\") pod \"telemetry-operator-controller-manager-58d5ff84df-6qvtl\" (UID: \"2536d7d4-e60f-44a6-8314-cfd34b5545e9\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.299875 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4gj5\" (UniqueName: \"kubernetes.io/projected/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-kube-api-access-g4gj5\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.307096 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.307102 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.308413 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.308501 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.308940 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwsxz\" (UniqueName: \"kubernetes.io/projected/59051068-3252-47e6-a826-79d1951c3d5c-kube-api-access-jwsxz\") pod \"placement-operator-controller-manager-78f8948974-sxfkp\" (UID: \"59051068-3252-47e6-a826-79d1951c3d5c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.309013 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.309054 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtwgt\" (UniqueName: \"kubernetes.io/projected/5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8-kube-api-access-jtwgt\") pod \"swift-operator-controller-manager-9d58d64bc-69wkl\" (UID: \"5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.309085 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-279hf\" (UniqueName: \"kubernetes.io/projected/2b5f3682-eb12-440c-a016-f165ec4e7cfa-kube-api-access-279hf\") pod \"ovn-operator-controller-manager-b6456fdb6-lr526\" (UID: \"2b5f3682-eb12-440c-a016-f165ec4e7cfa\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.310803 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.310905 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert podName:fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae nodeName:}" failed. No retries permitted until 2025-12-08 15:01:09.810884829 +0000 UTC m=+890.910890934 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4lcw4" (UID: "fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.330938 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-j98fz" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.334074 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.334638 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.335758 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.341071 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-z8xb4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.344642 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwsxz\" (UniqueName: \"kubernetes.io/projected/59051068-3252-47e6-a826-79d1951c3d5c-kube-api-access-jwsxz\") pod \"placement-operator-controller-manager-78f8948974-sxfkp\" (UID: \"59051068-3252-47e6-a826-79d1951c3d5c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.350270 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.369051 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4gj5\" (UniqueName: \"kubernetes.io/projected/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-kube-api-access-g4gj5\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.385994 4894 generic.go:334] "Generic (PLEG): container finished" podID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerID="cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d" exitCode=0 Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.386047 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerDied","Data":"cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d"} Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.389592 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-279hf\" (UniqueName: \"kubernetes.io/projected/2b5f3682-eb12-440c-a016-f165ec4e7cfa-kube-api-access-279hf\") pod \"ovn-operator-controller-manager-b6456fdb6-lr526\" (UID: \"2b5f3682-eb12-440c-a016-f165ec4e7cfa\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.411292 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtwgt\" (UniqueName: \"kubernetes.io/projected/5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8-kube-api-access-jtwgt\") pod \"swift-operator-controller-manager-9d58d64bc-69wkl\" (UID: \"5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.411410 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw5b6\" (UniqueName: \"kubernetes.io/projected/2536d7d4-e60f-44a6-8314-cfd34b5545e9-kube-api-access-gw5b6\") pod \"telemetry-operator-controller-manager-58d5ff84df-6qvtl\" (UID: \"2536d7d4-e60f-44a6-8314-cfd34b5545e9\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.415634 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.442098 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw5b6\" (UniqueName: \"kubernetes.io/projected/2536d7d4-e60f-44a6-8314-cfd34b5545e9-kube-api-access-gw5b6\") pod \"telemetry-operator-controller-manager-58d5ff84df-6qvtl\" (UID: \"2536d7d4-e60f-44a6-8314-cfd34b5545e9\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.447589 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtwgt\" (UniqueName: \"kubernetes.io/projected/5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8-kube-api-access-jtwgt\") pod \"swift-operator-controller-manager-9d58d64bc-69wkl\" (UID: \"5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.487332 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.488686 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.489596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.491305 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.503701 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.512381 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr22d\" (UniqueName: \"kubernetes.io/projected/8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76-kube-api-access-jr22d\") pod \"watcher-operator-controller-manager-667bd8d554-rnqzt\" (UID: \"8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.512438 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z7lm\" (UniqueName: \"kubernetes.io/projected/8b8eda01-4bc4-4eca-b909-901024ef466b-kube-api-access-8z7lm\") pod \"test-operator-controller-manager-5854674fcc-zhwdn\" (UID: \"8b8eda01-4bc4-4eca-b909-901024ef466b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.514693 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ffwrr" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.514947 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.516376 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.550216 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.551079 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.553210 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8kxcq" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.559392 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.619319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4jjm\" (UniqueName: \"kubernetes.io/projected/69d7f3da-48c5-47e8-853b-0cca6246d96e-kube-api-access-w4jjm\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.619387 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr22d\" (UniqueName: \"kubernetes.io/projected/8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76-kube-api-access-jr22d\") pod \"watcher-operator-controller-manager-667bd8d554-rnqzt\" (UID: \"8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.619518 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z7lm\" (UniqueName: \"kubernetes.io/projected/8b8eda01-4bc4-4eca-b909-901024ef466b-kube-api-access-8z7lm\") pod \"test-operator-controller-manager-5854674fcc-zhwdn\" (UID: \"8b8eda01-4bc4-4eca-b909-901024ef466b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.619568 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.619601 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.631033 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.633956 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.645146 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr22d\" (UniqueName: \"kubernetes.io/projected/8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76-kube-api-access-jr22d\") pod \"watcher-operator-controller-manager-667bd8d554-rnqzt\" (UID: \"8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.667260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z7lm\" (UniqueName: \"kubernetes.io/projected/8b8eda01-4bc4-4eca-b909-901024ef466b-kube-api-access-8z7lm\") pod \"test-operator-controller-manager-5854674fcc-zhwdn\" (UID: \"8b8eda01-4bc4-4eca-b909-901024ef466b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.714103 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.721638 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.721676 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.721755 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4jjm\" (UniqueName: \"kubernetes.io/projected/69d7f3da-48c5-47e8-853b-0cca6246d96e-kube-api-access-w4jjm\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.721780 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srgjs\" (UniqueName: \"kubernetes.io/projected/9c6f4113-6d2f-4fce-b76b-8802b3b605c6-kube-api-access-srgjs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-97mkf\" (UID: \"9c6f4113-6d2f-4fce-b76b-8802b3b605c6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.721937 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.721975 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:10.221962565 +0000 UTC m=+891.321968670 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.722189 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.722215 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:10.222208382 +0000 UTC m=+891.322214497 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "metrics-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.727993 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.784098 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4jjm\" (UniqueName: \"kubernetes.io/projected/69d7f3da-48c5-47e8-853b-0cca6246d96e-kube-api-access-w4jjm\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.823858 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srgjs\" (UniqueName: \"kubernetes.io/projected/9c6f4113-6d2f-4fce-b76b-8802b3b605c6-kube-api-access-srgjs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-97mkf\" (UID: \"9c6f4113-6d2f-4fce-b76b-8802b3b605c6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.823982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.824152 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: E1208 15:01:09.824212 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert podName:fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae nodeName:}" failed. No retries permitted until 2025-12-08 15:01:10.824193674 +0000 UTC m=+891.924199789 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4lcw4" (UID: "fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.857956 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srgjs\" (UniqueName: \"kubernetes.io/projected/9c6f4113-6d2f-4fce-b76b-8802b3b605c6-kube-api-access-srgjs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-97mkf\" (UID: \"9c6f4113-6d2f-4fce-b76b-8802b3b605c6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.860550 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc"] Dec 08 15:01:09 crc kubenswrapper[4894]: I1208 15:01:09.928748 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.135278 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.136081 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.136160 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert podName:c3e89da4-2a46-4982-9392-b0cf87a1edc3 nodeName:}" failed. No retries permitted until 2025-12-08 15:01:12.136138408 +0000 UTC m=+893.236144523 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert") pod "infra-operator-controller-manager-78d48bff9d-4xrts" (UID: "c3e89da4-2a46-4982-9392-b0cf87a1edc3") : secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.244401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.244613 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.244542 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.244693 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:11.244674435 +0000 UTC m=+892.344680550 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "webhook-server-cert" not found Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.244779 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.244829 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:11.24480515 +0000 UTC m=+892.344811255 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "metrics-server-cert" not found Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.340560 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb"] Dec 08 15:01:10 crc kubenswrapper[4894]: W1208 15:01:10.350659 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafda731d_73a8_443f_9647_a774ca8c3146.slice/crio-46d36e0fce7b87657192169d14af89671f374a261f878e7e9ce0696b30bad9ee WatchSource:0}: Error finding container 46d36e0fce7b87657192169d14af89671f374a261f878e7e9ce0696b30bad9ee: Status 404 returned error can't find the container with id 46d36e0fce7b87657192169d14af89671f374a261f878e7e9ce0696b30bad9ee Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.352400 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-47bpp"] Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.396572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" event={"ID":"f7ec7a43-f5b5-474f-af15-d0e664699477","Type":"ContainerStarted","Data":"f16daafcdc3cdd09b7b88f70bfbd9c9aace1ac292e59e8f3817a0ee525582808"} Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.397361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" event={"ID":"28cfecb9-9c7c-49ae-af33-fc2d9acd374a","Type":"ContainerStarted","Data":"6b2b9a45b385e3049067089da7274279e67b7952a69b3be0166ec3407e45d472"} Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.398083 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" event={"ID":"afda731d-73a8-443f-9647-a774ca8c3146","Type":"ContainerStarted","Data":"46d36e0fce7b87657192169d14af89671f374a261f878e7e9ce0696b30bad9ee"} Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.398708 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" event={"ID":"e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72","Type":"ContainerStarted","Data":"348ef08f023a07942114cb17482922d5da104cf68850ea96e5935d01fe20d826"} Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.581422 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj"] Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.600569 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p"] Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.609177 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c"] Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.624800 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4"] Dec 08 15:01:10 crc kubenswrapper[4894]: W1208 15:01:10.625255 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6585a6c_b50d_41d3_a080_ab7cb27f9199.slice/crio-3085eb70866c0817a52f280f5ca128d113809447ad666901556dc83ba6301787 WatchSource:0}: Error finding container 3085eb70866c0817a52f280f5ca128d113809447ad666901556dc83ba6301787: Status 404 returned error can't find the container with id 3085eb70866c0817a52f280f5ca128d113809447ad666901556dc83ba6301787 Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.648514 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw"] Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.655622 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt"] Dec 08 15:01:10 crc kubenswrapper[4894]: W1208 15:01:10.659724 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d7b9bb8_c977_45cb_b198_5f047059a0a4.slice/crio-ba41692084407ed01a5bdd78e6690657363fdffde0524220b18cdb66655f1251 WatchSource:0}: Error finding container ba41692084407ed01a5bdd78e6690657363fdffde0524220b18cdb66655f1251: Status 404 returned error can't find the container with id ba41692084407ed01a5bdd78e6690657363fdffde0524220b18cdb66655f1251 Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.660761 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t"] Dec 08 15:01:10 crc kubenswrapper[4894]: W1208 15:01:10.662472 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2df93571_3b4e_4ab0_ad8b_34d2ef76034c.slice/crio-7c0031379635b4ba8b9c041128af7d4ddf537681ab3dd8edf304b20d56f0a876 WatchSource:0}: Error finding container 7c0031379635b4ba8b9c041128af7d4ddf537681ab3dd8edf304b20d56f0a876: Status 404 returned error can't find the container with id 7c0031379635b4ba8b9c041128af7d4ddf537681ab3dd8edf304b20d56f0a876 Dec 08 15:01:10 crc kubenswrapper[4894]: W1208 15:01:10.665142 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8a908e6_3df2_4dc9_9c9c_068e91dc839c.slice/crio-a06216d3d19863c45f7f404461798b4b7ec5ea23f5198ce1c06c0754d350996b WatchSource:0}: Error finding container a06216d3d19863c45f7f404461798b4b7ec5ea23f5198ce1c06c0754d350996b: Status 404 returned error can't find the container with id a06216d3d19863c45f7f404461798b4b7ec5ea23f5198ce1c06c0754d350996b Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.790525 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp"] Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.795569 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb"] Dec 08 15:01:10 crc kubenswrapper[4894]: W1208 15:01:10.807208 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59051068_3252_47e6_a826_79d1951c3d5c.slice/crio-6748c0864a50daeb123a530267c01756b99df714a15afe6c3b46ea150728d396 WatchSource:0}: Error finding container 6748c0864a50daeb123a530267c01756b99df714a15afe6c3b46ea150728d396: Status 404 returned error can't find the container with id 6748c0864a50daeb123a530267c01756b99df714a15afe6c3b46ea150728d396 Dec 08 15:01:10 crc kubenswrapper[4894]: W1208 15:01:10.808012 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc91237dd_8bb9_4240_b24a_e19b58687d2e.slice/crio-a1eb48480b60cbf7d3fb6fcde684b18aeebbfe063eeeb9cb317eb6efcf286fc9 WatchSource:0}: Error finding container a1eb48480b60cbf7d3fb6fcde684b18aeebbfe063eeeb9cb317eb6efcf286fc9: Status 404 returned error can't find the container with id a1eb48480b60cbf7d3fb6fcde684b18aeebbfe063eeeb9cb317eb6efcf286fc9 Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.832260 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf"] Dec 08 15:01:10 crc kubenswrapper[4894]: I1208 15:01:10.855510 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.855670 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:10 crc kubenswrapper[4894]: E1208 15:01:10.855725 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert podName:fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae nodeName:}" failed. No retries permitted until 2025-12-08 15:01:12.855711248 +0000 UTC m=+893.955717363 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4lcw4" (UID: "fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.013869 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd"] Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.021532 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl"] Dec 08 15:01:11 crc kubenswrapper[4894]: W1208 15:01:11.022616 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb804bba_5a84_4d23_93ca_e7118afc2af8.slice/crio-2ba55e295051bc6471e63aa3caf4b73caf632f685ae1517ad37c9fd7163154e6 WatchSource:0}: Error finding container 2ba55e295051bc6471e63aa3caf4b73caf632f685ae1517ad37c9fd7163154e6: Status 404 returned error can't find the container with id 2ba55e295051bc6471e63aa3caf4b73caf632f685ae1517ad37c9fd7163154e6 Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.037807 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526"] Dec 08 15:01:11 crc kubenswrapper[4894]: W1208 15:01:11.042000 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ed01f8d_08b6_4fdf_8478_0c0c373ff3a8.slice/crio-a78b920527b6d8fdb3332737245bd7c36b01f7bb196fae1c8b06f6f984eb4719 WatchSource:0}: Error finding container a78b920527b6d8fdb3332737245bd7c36b01f7bb196fae1c8b06f6f984eb4719: Status 404 returned error can't find the container with id a78b920527b6d8fdb3332737245bd7c36b01f7bb196fae1c8b06f6f984eb4719 Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.044065 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jtwgt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-69wkl_openstack-operators(5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.053914 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jtwgt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-69wkl_openstack-operators(5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.054108 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt"] Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.056420 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" podUID="5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8" Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.060845 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl"] Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.062202 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-279hf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-lr526_openstack-operators(2b5f3682-eb12-440c-a016-f165ec4e7cfa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.064785 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn"] Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.069425 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-279hf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-lr526_openstack-operators(2b5f3682-eb12-440c-a016-f165ec4e7cfa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.070742 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" podUID="2b5f3682-eb12-440c-a016-f165ec4e7cfa" Dec 08 15:01:11 crc kubenswrapper[4894]: W1208 15:01:11.072792 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b8eda01_4bc4_4eca_b909_901024ef466b.slice/crio-4a82d151ffbb6500c5f0b67cb5494d8366e8436e0c67a26cc67fea50fb0a720f WatchSource:0}: Error finding container 4a82d151ffbb6500c5f0b67cb5494d8366e8436e0c67a26cc67fea50fb0a720f: Status 404 returned error can't find the container with id 4a82d151ffbb6500c5f0b67cb5494d8366e8436e0c67a26cc67fea50fb0a720f Dec 08 15:01:11 crc kubenswrapper[4894]: W1208 15:01:11.074401 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f07e44c_34c8_4ff9_ba05_8ea4b46dfa76.slice/crio-5b7b7c3ce1ff20858216737a2b7535de17e369570d2390b0aebeb161e66fcbf5 WatchSource:0}: Error finding container 5b7b7c3ce1ff20858216737a2b7535de17e369570d2390b0aebeb161e66fcbf5: Status 404 returned error can't find the container with id 5b7b7c3ce1ff20858216737a2b7535de17e369570d2390b0aebeb161e66fcbf5 Dec 08 15:01:11 crc kubenswrapper[4894]: W1208 15:01:11.076709 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2536d7d4_e60f_44a6_8314_cfd34b5545e9.slice/crio-7bb020649d1409bb26cc209461945526960683e1ae5617d47c0a49b78a2a199e WatchSource:0}: Error finding container 7bb020649d1409bb26cc209461945526960683e1ae5617d47c0a49b78a2a199e: Status 404 returned error can't find the container with id 7bb020649d1409bb26cc209461945526960683e1ae5617d47c0a49b78a2a199e Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.077958 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8z7lm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-zhwdn_openstack-operators(8b8eda01-4bc4-4eca-b909-901024ef466b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.078060 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jr22d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-rnqzt_openstack-operators(8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.091461 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jr22d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-rnqzt_openstack-operators(8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.092671 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" podUID="8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.093058 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8z7lm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-zhwdn_openstack-operators(8b8eda01-4bc4-4eca-b909-901024ef466b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.094315 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" podUID="8b8eda01-4bc4-4eca-b909-901024ef466b" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.095004 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gw5b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-6qvtl_openstack-operators(2536d7d4-e60f-44a6-8314-cfd34b5545e9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.097689 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gw5b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-6qvtl_openstack-operators(2536d7d4-e60f-44a6-8314-cfd34b5545e9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.099302 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" podUID="2536d7d4-e60f-44a6-8314-cfd34b5545e9" Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.273458 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.273602 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.273663 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:13.27364653 +0000 UTC m=+894.373652655 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "metrics-server-cert" not found Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.273613 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.273664 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.273729 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:13.273720722 +0000 UTC m=+894.373726837 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "webhook-server-cert" not found Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.412081 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" event={"ID":"8b8eda01-4bc4-4eca-b909-901024ef466b","Type":"ContainerStarted","Data":"4a82d151ffbb6500c5f0b67cb5494d8366e8436e0c67a26cc67fea50fb0a720f"} Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.416197 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" podUID="8b8eda01-4bc4-4eca-b909-901024ef466b" Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.417019 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerStarted","Data":"a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.422477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" event={"ID":"db804bba-5a84-4d23-93ca-e7118afc2af8","Type":"ContainerStarted","Data":"2ba55e295051bc6471e63aa3caf4b73caf632f685ae1517ad37c9fd7163154e6"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.427768 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" event={"ID":"aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2","Type":"ContainerStarted","Data":"f0deffecda137eceef2f17cad94392bb2f8f6d3b6796195b51a2188fd3ca13ee"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.436431 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" event={"ID":"750c567a-693a-401b-8e7f-d8f1befbeacb","Type":"ContainerStarted","Data":"41ab3ab6b49267812a1085be53e32f63dc8947339a9687c22aec8ffa65f7be0c"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.438845 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" event={"ID":"2b5f3682-eb12-440c-a016-f165ec4e7cfa","Type":"ContainerStarted","Data":"26f78e5e4b1dfd7ae8509c33f62c9c49353dce59de9e920010ec6f69accdb826"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.441213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" event={"ID":"8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76","Type":"ContainerStarted","Data":"5b7b7c3ce1ff20858216737a2b7535de17e369570d2390b0aebeb161e66fcbf5"} Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.442263 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" podUID="2b5f3682-eb12-440c-a016-f165ec4e7cfa" Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.443389 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" podUID="8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76" Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.443770 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" event={"ID":"c91237dd-8bb9-4240-b24a-e19b58687d2e","Type":"ContainerStarted","Data":"a1eb48480b60cbf7d3fb6fcde684b18aeebbfe063eeeb9cb317eb6efcf286fc9"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.446893 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" event={"ID":"2536d7d4-e60f-44a6-8314-cfd34b5545e9","Type":"ContainerStarted","Data":"7bb020649d1409bb26cc209461945526960683e1ae5617d47c0a49b78a2a199e"} Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.449398 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" podUID="2536d7d4-e60f-44a6-8314-cfd34b5545e9" Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.457035 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" event={"ID":"4ee4db87-8c97-4002-844b-6c55e1a41b2e","Type":"ContainerStarted","Data":"02a594a7698fdf0ef9c8aff1d4802a1e47fb60ad64a88c898c20450280ade7a7"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.461618 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" event={"ID":"59051068-3252-47e6-a826-79d1951c3d5c","Type":"ContainerStarted","Data":"6748c0864a50daeb123a530267c01756b99df714a15afe6c3b46ea150728d396"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.463845 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" event={"ID":"e6585a6c-b50d-41d3-a080-ab7cb27f9199","Type":"ContainerStarted","Data":"3085eb70866c0817a52f280f5ca128d113809447ad666901556dc83ba6301787"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.465848 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" event={"ID":"5d7b9bb8-c977-45cb-b198-5f047059a0a4","Type":"ContainerStarted","Data":"ba41692084407ed01a5bdd78e6690657363fdffde0524220b18cdb66655f1251"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.471644 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" event={"ID":"b8a908e6-3df2-4dc9-9c9c-068e91dc839c","Type":"ContainerStarted","Data":"a06216d3d19863c45f7f404461798b4b7ec5ea23f5198ce1c06c0754d350996b"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.479932 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" event={"ID":"2df93571-3b4e-4ab0-ad8b-34d2ef76034c","Type":"ContainerStarted","Data":"7c0031379635b4ba8b9c041128af7d4ddf537681ab3dd8edf304b20d56f0a876"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.484748 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" event={"ID":"5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8","Type":"ContainerStarted","Data":"a78b920527b6d8fdb3332737245bd7c36b01f7bb196fae1c8b06f6f984eb4719"} Dec 08 15:01:11 crc kubenswrapper[4894]: I1208 15:01:11.495974 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" event={"ID":"9c6f4113-6d2f-4fce-b76b-8802b3b605c6","Type":"ContainerStarted","Data":"a0ec1837b4adaa26812658795ba6bad9891dd320d1f93c894c861575a9be9f55"} Dec 08 15:01:11 crc kubenswrapper[4894]: E1208 15:01:11.507621 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" podUID="5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8" Dec 08 15:01:12 crc kubenswrapper[4894]: I1208 15:01:12.210687 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.211216 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.211266 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert podName:c3e89da4-2a46-4982-9392-b0cf87a1edc3 nodeName:}" failed. No retries permitted until 2025-12-08 15:01:16.211250036 +0000 UTC m=+897.311256151 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert") pod "infra-operator-controller-manager-78d48bff9d-4xrts" (UID: "c3e89da4-2a46-4982-9392-b0cf87a1edc3") : secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.538654 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" podUID="5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8" Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.543727 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" podUID="2536d7d4-e60f-44a6-8314-cfd34b5545e9" Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.543807 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" podUID="2b5f3682-eb12-440c-a016-f165ec4e7cfa" Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.544231 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" podUID="8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76" Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.548263 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" podUID="8b8eda01-4bc4-4eca-b909-901024ef466b" Dec 08 15:01:12 crc kubenswrapper[4894]: I1208 15:01:12.931091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.931380 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:12 crc kubenswrapper[4894]: E1208 15:01:12.931425 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert podName:fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae nodeName:}" failed. No retries permitted until 2025-12-08 15:01:16.931411114 +0000 UTC m=+898.031417229 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4lcw4" (UID: "fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:13 crc kubenswrapper[4894]: I1208 15:01:13.344386 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:13 crc kubenswrapper[4894]: I1208 15:01:13.344460 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:13 crc kubenswrapper[4894]: E1208 15:01:13.344711 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 15:01:13 crc kubenswrapper[4894]: E1208 15:01:13.344781 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:17.344762742 +0000 UTC m=+898.444768857 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "metrics-server-cert" not found Dec 08 15:01:13 crc kubenswrapper[4894]: E1208 15:01:13.344875 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 15:01:13 crc kubenswrapper[4894]: E1208 15:01:13.344910 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:17.344901416 +0000 UTC m=+898.444907541 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "webhook-server-cert" not found Dec 08 15:01:13 crc kubenswrapper[4894]: I1208 15:01:13.606466 4894 generic.go:334] "Generic (PLEG): container finished" podID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerID="a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455" exitCode=0 Dec 08 15:01:13 crc kubenswrapper[4894]: I1208 15:01:13.606510 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerDied","Data":"a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455"} Dec 08 15:01:16 crc kubenswrapper[4894]: I1208 15:01:16.293346 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:16 crc kubenswrapper[4894]: E1208 15:01:16.294094 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:16 crc kubenswrapper[4894]: E1208 15:01:16.294147 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert podName:c3e89da4-2a46-4982-9392-b0cf87a1edc3 nodeName:}" failed. No retries permitted until 2025-12-08 15:01:24.294133462 +0000 UTC m=+905.394139577 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert") pod "infra-operator-controller-manager-78d48bff9d-4xrts" (UID: "c3e89da4-2a46-4982-9392-b0cf87a1edc3") : secret "infra-operator-webhook-server-cert" not found Dec 08 15:01:17 crc kubenswrapper[4894]: I1208 15:01:17.004636 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:17 crc kubenswrapper[4894]: E1208 15:01:17.004973 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:17 crc kubenswrapper[4894]: E1208 15:01:17.005032 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert podName:fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae nodeName:}" failed. No retries permitted until 2025-12-08 15:01:25.005015519 +0000 UTC m=+906.105021624 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert") pod "openstack-baremetal-operator-controller-manager-84b575879f4lcw4" (UID: "fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 08 15:01:17 crc kubenswrapper[4894]: I1208 15:01:17.412974 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:17 crc kubenswrapper[4894]: I1208 15:01:17.413028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:17 crc kubenswrapper[4894]: E1208 15:01:17.413197 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 08 15:01:17 crc kubenswrapper[4894]: E1208 15:01:17.413200 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 15:01:17 crc kubenswrapper[4894]: E1208 15:01:17.413245 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:25.413231054 +0000 UTC m=+906.513237169 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "metrics-server-cert" not found Dec 08 15:01:17 crc kubenswrapper[4894]: E1208 15:01:17.413277 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:25.413259695 +0000 UTC m=+906.513265810 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "webhook-server-cert" not found Dec 08 15:01:24 crc kubenswrapper[4894]: I1208 15:01:24.322061 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:24 crc kubenswrapper[4894]: I1208 15:01:24.328909 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3e89da4-2a46-4982-9392-b0cf87a1edc3-cert\") pod \"infra-operator-controller-manager-78d48bff9d-4xrts\" (UID: \"c3e89da4-2a46-4982-9392-b0cf87a1edc3\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:24 crc kubenswrapper[4894]: I1208 15:01:24.435690 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-964zj" Dec 08 15:01:24 crc kubenswrapper[4894]: I1208 15:01:24.444066 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:25 crc kubenswrapper[4894]: I1208 15:01:25.032149 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:25 crc kubenswrapper[4894]: I1208 15:01:25.037636 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879f4lcw4\" (UID: \"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:25 crc kubenswrapper[4894]: I1208 15:01:25.055225 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5gjs4" Dec 08 15:01:25 crc kubenswrapper[4894]: I1208 15:01:25.064340 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:25 crc kubenswrapper[4894]: E1208 15:01:25.155270 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 08 15:01:25 crc kubenswrapper[4894]: E1208 15:01:25.155734 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dw67p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-mtkrb_openstack-operators(afda731d-73a8-443f-9647-a774ca8c3146): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:25 crc kubenswrapper[4894]: I1208 15:01:25.438208 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:25 crc kubenswrapper[4894]: I1208 15:01:25.438346 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:25 crc kubenswrapper[4894]: E1208 15:01:25.438377 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 08 15:01:25 crc kubenswrapper[4894]: E1208 15:01:25.438437 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs podName:69d7f3da-48c5-47e8-853b-0cca6246d96e nodeName:}" failed. No retries permitted until 2025-12-08 15:01:41.438417302 +0000 UTC m=+922.538423407 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs") pod "openstack-operator-controller-manager-8695b4b66-c64t7" (UID: "69d7f3da-48c5-47e8-853b-0cca6246d96e") : secret "webhook-server-cert" not found Dec 08 15:01:25 crc kubenswrapper[4894]: I1208 15:01:25.456630 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-metrics-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:25 crc kubenswrapper[4894]: E1208 15:01:25.996241 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 08 15:01:25 crc kubenswrapper[4894]: E1208 15:01:25.996663 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7c6hp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-w7xbj_openstack-operators(aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:32 crc kubenswrapper[4894]: E1208 15:01:32.136855 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 08 15:01:32 crc kubenswrapper[4894]: E1208 15:01:32.137542 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhfdk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-cqq4p_openstack-operators(750c567a-693a-401b-8e7f-d8f1befbeacb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:32 crc kubenswrapper[4894]: E1208 15:01:32.525072 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 08 15:01:32 crc kubenswrapper[4894]: E1208 15:01:32.525278 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2xbq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-kh5v9_openstack-operators(28cfecb9-9c7c-49ae-af33-fc2d9acd374a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:33 crc kubenswrapper[4894]: E1208 15:01:33.657633 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 08 15:01:33 crc kubenswrapper[4894]: E1208 15:01:33.658028 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jwsxz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-sxfkp_openstack-operators(59051068-3252-47e6-a826-79d1951c3d5c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:34 crc kubenswrapper[4894]: E1208 15:01:34.542056 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 08 15:01:34 crc kubenswrapper[4894]: E1208 15:01:34.542257 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gcphm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-47bpp_openstack-operators(f7ec7a43-f5b5-474f-af15-d0e664699477): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:35 crc kubenswrapper[4894]: E1208 15:01:35.933649 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 08 15:01:35 crc kubenswrapper[4894]: E1208 15:01:35.933859 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8jvwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-2sfqd_openstack-operators(db804bba-5a84-4d23-93ca-e7118afc2af8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:37 crc kubenswrapper[4894]: I1208 15:01:37.296359 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:01:37 crc kubenswrapper[4894]: I1208 15:01:37.296666 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:01:37 crc kubenswrapper[4894]: E1208 15:01:37.631880 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 08 15:01:37 crc kubenswrapper[4894]: E1208 15:01:37.632166 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9dtln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-lbszt_openstack-operators(2df93571-3b4e-4ab0-ad8b-34d2ef76034c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:41 crc kubenswrapper[4894]: I1208 15:01:41.484558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:41 crc kubenswrapper[4894]: I1208 15:01:41.493321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/69d7f3da-48c5-47e8-853b-0cca6246d96e-webhook-certs\") pod \"openstack-operator-controller-manager-8695b4b66-c64t7\" (UID: \"69d7f3da-48c5-47e8-853b-0cca6246d96e\") " pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:41 crc kubenswrapper[4894]: I1208 15:01:41.729135 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ffwrr" Dec 08 15:01:41 crc kubenswrapper[4894]: I1208 15:01:41.738726 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:42 crc kubenswrapper[4894]: E1208 15:01:42.279329 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 08 15:01:42 crc kubenswrapper[4894]: E1208 15:01:42.279884 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7bvvz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-jc2rb_openstack-operators(c91237dd-8bb9-4240-b24a-e19b58687d2e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:42 crc kubenswrapper[4894]: E1208 15:01:42.848401 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 08 15:01:42 crc kubenswrapper[4894]: E1208 15:01:42.848582 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-srgjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-97mkf_openstack-operators(9c6f4113-6d2f-4fce-b76b-8802b3b605c6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:01:42 crc kubenswrapper[4894]: E1208 15:01:42.850261 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" podUID="9c6f4113-6d2f-4fce-b76b-8802b3b605c6" Dec 08 15:01:43 crc kubenswrapper[4894]: I1208 15:01:43.729943 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts"] Dec 08 15:01:43 crc kubenswrapper[4894]: I1208 15:01:43.744432 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4"] Dec 08 15:01:43 crc kubenswrapper[4894]: E1208 15:01:43.857705 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" podUID="9c6f4113-6d2f-4fce-b76b-8802b3b605c6" Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.556243 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7"] Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.870035 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" event={"ID":"5d7b9bb8-c977-45cb-b198-5f047059a0a4","Type":"ContainerStarted","Data":"bf05fcabb917ec9aa419fceec4c9b4ee37e93bb46ffd16b1ed9d1ce034b1d5db"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.872272 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" event={"ID":"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae","Type":"ContainerStarted","Data":"dfc9df2b734175c0ec3691f74a5bc09a0b624580af60ac34ca2abe8a475d1a1c"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.873598 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" event={"ID":"c3e89da4-2a46-4982-9392-b0cf87a1edc3","Type":"ContainerStarted","Data":"0c331aeb69efa178374d52f12c11aa1e89f6a1806c41a50264adc6b87e5665cd"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.878269 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" event={"ID":"e6585a6c-b50d-41d3-a080-ab7cb27f9199","Type":"ContainerStarted","Data":"d1316dc98c534e10b081e8b9d174f92541fd7635f96f94d93f42370221139a0e"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.879844 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" event={"ID":"e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72","Type":"ContainerStarted","Data":"959fe3f468c6b62a943a3b4ebcde22d1f5a993d5b53531bc20f4b5b51263fd62"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.883483 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerStarted","Data":"20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.897184 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" event={"ID":"4ee4db87-8c97-4002-844b-6c55e1a41b2e","Type":"ContainerStarted","Data":"4c1127fa6b0357aef45d16c6305590c46ab96382066726ab0d4cbb34b8c755ec"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.902148 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" event={"ID":"b8a908e6-3df2-4dc9-9c9c-068e91dc839c","Type":"ContainerStarted","Data":"31b0741c0988857e53576e930698bb269d737c303e53f037389fe96f9c1dbeb9"} Dec 08 15:01:44 crc kubenswrapper[4894]: I1208 15:01:44.907162 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r6hn7" podStartSLOduration=4.47311436 podStartE2EDuration="37.907147714s" podCreationTimestamp="2025-12-08 15:01:07 +0000 UTC" firstStartedPulling="2025-12-08 15:01:09.41794468 +0000 UTC m=+890.517950795" lastFinishedPulling="2025-12-08 15:01:42.851978044 +0000 UTC m=+923.951984149" observedRunningTime="2025-12-08 15:01:44.90543986 +0000 UTC m=+926.005445965" watchObservedRunningTime="2025-12-08 15:01:44.907147714 +0000 UTC m=+926.007153829" Dec 08 15:01:45 crc kubenswrapper[4894]: W1208 15:01:45.444630 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69d7f3da_48c5_47e8_853b_0cca6246d96e.slice/crio-eaea11919c2a120adeff562200f20827e4bc541c9a98f1189e5cc20f9c37a27d WatchSource:0}: Error finding container eaea11919c2a120adeff562200f20827e4bc541c9a98f1189e5cc20f9c37a27d: Status 404 returned error can't find the container with id eaea11919c2a120adeff562200f20827e4bc541c9a98f1189e5cc20f9c37a27d Dec 08 15:01:45 crc kubenswrapper[4894]: I1208 15:01:45.935677 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" event={"ID":"2b5f3682-eb12-440c-a016-f165ec4e7cfa","Type":"ContainerStarted","Data":"71ef11855168ebf693305638f9965feb1ddad38d228681fdd42f89b8ce97ba13"} Dec 08 15:01:46 crc kubenswrapper[4894]: I1208 15:01:46.053096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" event={"ID":"69d7f3da-48c5-47e8-853b-0cca6246d96e","Type":"ContainerStarted","Data":"eaea11919c2a120adeff562200f20827e4bc541c9a98f1189e5cc20f9c37a27d"} Dec 08 15:01:46 crc kubenswrapper[4894]: I1208 15:01:46.061017 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" event={"ID":"8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76","Type":"ContainerStarted","Data":"47afe6bc4acc2a54dd0b547eeb7ab8b66a52f22bf03e00b7a868a9decb21c371"} Dec 08 15:01:47 crc kubenswrapper[4894]: I1208 15:01:47.750209 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:47 crc kubenswrapper[4894]: I1208 15:01:47.750278 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:48 crc kubenswrapper[4894]: I1208 15:01:48.814894 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r6hn7" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="registry-server" probeResult="failure" output=< Dec 08 15:01:48 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 08 15:01:48 crc kubenswrapper[4894]: > Dec 08 15:01:50 crc kubenswrapper[4894]: E1208 15:01:50.584923 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" podUID="aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2" Dec 08 15:01:50 crc kubenswrapper[4894]: E1208 15:01:50.725716 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" podUID="2df93571-3b4e-4ab0-ad8b-34d2ef76034c" Dec 08 15:01:50 crc kubenswrapper[4894]: E1208 15:01:50.836193 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" podUID="f7ec7a43-f5b5-474f-af15-d0e664699477" Dec 08 15:01:50 crc kubenswrapper[4894]: E1208 15:01:50.848888 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" podUID="750c567a-693a-401b-8e7f-d8f1befbeacb" Dec 08 15:01:50 crc kubenswrapper[4894]: E1208 15:01:50.861419 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" podUID="59051068-3252-47e6-a826-79d1951c3d5c" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.122402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" event={"ID":"2536d7d4-e60f-44a6-8314-cfd34b5545e9","Type":"ContainerStarted","Data":"3017f25263591f504a97be011dc2351d63cb5e0f7db9a4e0b5d088d14a969db5"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.133885 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" event={"ID":"f7ec7a43-f5b5-474f-af15-d0e664699477","Type":"ContainerStarted","Data":"4b0d65a2a0a800e26a49ff18d725d9687e9ddb873a10eabdeaf378215d14109f"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.140714 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:01:51 crc kubenswrapper[4894]: E1208 15:01:51.155163 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" podUID="db804bba-5a84-4d23-93ca-e7118afc2af8" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.173242 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" event={"ID":"b8a908e6-3df2-4dc9-9c9c-068e91dc839c","Type":"ContainerStarted","Data":"e9898f0a5ec12b5917911db1ccc4e7de3b1bf9a786a90278e7404e984880f6a2"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.174486 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.187572 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.235367 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-ffvpw" podStartSLOduration=3.83855332 podStartE2EDuration="43.235345479s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.95959194 +0000 UTC m=+892.059598055" lastFinishedPulling="2025-12-08 15:01:50.356384099 +0000 UTC m=+931.456390214" observedRunningTime="2025-12-08 15:01:51.228151083 +0000 UTC m=+932.328157198" watchObservedRunningTime="2025-12-08 15:01:51.235345479 +0000 UTC m=+932.335351604" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.371103 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" event={"ID":"2df93571-3b4e-4ab0-ad8b-34d2ef76034c","Type":"ContainerStarted","Data":"575cec79ada4257ebf5a5d3d2285947a8619d226077b299396b45362a9d36f8a"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.371148 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" event={"ID":"c3e89da4-2a46-4982-9392-b0cf87a1edc3","Type":"ContainerStarted","Data":"b7e9ef46b3c10ffdd4a0aa1168c67f37761eefa701554bd28a35aecca25c8298"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.371173 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" event={"ID":"59051068-3252-47e6-a826-79d1951c3d5c","Type":"ContainerStarted","Data":"256a318d2bbbdb9ec81b02ece7e79b2329ab08083874f829e624eb50aa961ef1"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.371281 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" event={"ID":"5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8","Type":"ContainerStarted","Data":"6aab50ad66acc2bcab18373ca005af443e4e31f391d95949320f7aa07c813aa9"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.393331 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" event={"ID":"aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2","Type":"ContainerStarted","Data":"21d8046e0b4fd2ee8e73baf7f23aa37a0934feb141dfb447b81555a62b1f94cd"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.395548 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" event={"ID":"750c567a-693a-401b-8e7f-d8f1befbeacb","Type":"ContainerStarted","Data":"2df8e82fcb889f86fb517089cd20fa7c32af28d82ebe11eeaf549c0a05efcec2"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.405742 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" event={"ID":"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae","Type":"ContainerStarted","Data":"1b72a92a61f70220f5b5812c897cc72b7c6731c13c0603a5f9fa5a294a1db557"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.444719 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" event={"ID":"2b5f3682-eb12-440c-a016-f165ec4e7cfa","Type":"ContainerStarted","Data":"2f2e39275161f1e8f2bb68d07e629a2206516502b6622f34681e894a3c7c7c29"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.445786 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.450359 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" event={"ID":"8b8eda01-4bc4-4eca-b909-901024ef466b","Type":"ContainerStarted","Data":"94c3e60faf89a0f1ff7936634363da97ec227c4691fe0e60af1fa6b71d9011e2"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.451123 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.477079 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" event={"ID":"69d7f3da-48c5-47e8-853b-0cca6246d96e","Type":"ContainerStarted","Data":"8530b637a47593b61ce50f194b4169d88b59bd7e6fe92dd64aae25293a169ab0"} Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.478111 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:01:51 crc kubenswrapper[4894]: I1208 15:01:51.632404 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-lr526" podStartSLOduration=4.375138347 podStartE2EDuration="43.632388523s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:11.062061866 +0000 UTC m=+892.162067981" lastFinishedPulling="2025-12-08 15:01:50.319312042 +0000 UTC m=+931.419318157" observedRunningTime="2025-12-08 15:01:51.582054577 +0000 UTC m=+932.682060692" watchObservedRunningTime="2025-12-08 15:01:51.632388523 +0000 UTC m=+932.732394638" Dec 08 15:01:51 crc kubenswrapper[4894]: E1208 15:01:51.636170 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" podUID="c91237dd-8bb9-4240-b24a-e19b58687d2e" Dec 08 15:01:51 crc kubenswrapper[4894]: E1208 15:01:51.992664 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" podUID="afda731d-73a8-443f-9647-a774ca8c3146" Dec 08 15:01:52 crc kubenswrapper[4894]: E1208 15:01:52.039168 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" podUID="28cfecb9-9c7c-49ae-af33-fc2d9acd374a" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.507701 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" event={"ID":"aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2","Type":"ContainerStarted","Data":"589457c0ba7969642183f28d9cb56b666e5fa9fd6b81299c31c3e3ca56c6490f"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.508158 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.511009 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" event={"ID":"afda731d-73a8-443f-9647-a774ca8c3146","Type":"ContainerStarted","Data":"554fd9ba9ea3ed39d3593e9079abe93f8960e940ef241a3f5b6b158647141217"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.523108 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" event={"ID":"c3e89da4-2a46-4982-9392-b0cf87a1edc3","Type":"ContainerStarted","Data":"f9aac20dff850f6d69948c9b5afd44ff0bd17e8e2daf8cdad82ed080ad901496"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.523511 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.537896 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" event={"ID":"28cfecb9-9c7c-49ae-af33-fc2d9acd374a","Type":"ContainerStarted","Data":"46bd66bde214ef0923773113b9f56fa668f88e50550754b91a7de38b0786e0d2"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.543086 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" podStartSLOduration=3.486998518 podStartE2EDuration="44.543067582s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.959544258 +0000 UTC m=+892.059550383" lastFinishedPulling="2025-12-08 15:01:52.015613332 +0000 UTC m=+933.115619447" observedRunningTime="2025-12-08 15:01:52.537865638 +0000 UTC m=+933.637871753" watchObservedRunningTime="2025-12-08 15:01:52.543067582 +0000 UTC m=+933.643073697" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.545460 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" podStartSLOduration=43.545447546 podStartE2EDuration="43.545447546s" podCreationTimestamp="2025-12-08 15:01:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:01:51.636386999 +0000 UTC m=+932.736393114" watchObservedRunningTime="2025-12-08 15:01:52.545447546 +0000 UTC m=+933.645453661" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.582174 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" event={"ID":"e6585a6c-b50d-41d3-a080-ab7cb27f9199","Type":"ContainerStarted","Data":"0b8e89bcd5ade6d7ee37b917ea62e4217565d9e6fdbc3ba4f13ece3d4c7ad070"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.583355 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.593478 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.613966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" event={"ID":"2536d7d4-e60f-44a6-8314-cfd34b5545e9","Type":"ContainerStarted","Data":"de4887ceb983f466d7502f65c142000fd56cf748270531ebdc742d0451174c4e"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.614659 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.617188 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" event={"ID":"5d7b9bb8-c977-45cb-b198-5f047059a0a4","Type":"ContainerStarted","Data":"d516422652a6d8eedd9a6d9cb4fbbd056627e1ec73e3b75fd4a6b10eee92314b"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.617879 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.624529 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" podStartSLOduration=38.527381598 podStartE2EDuration="44.624511556s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:44.080281905 +0000 UTC m=+925.180288020" lastFinishedPulling="2025-12-08 15:01:50.177411863 +0000 UTC m=+931.277417978" observedRunningTime="2025-12-08 15:01:52.622605397 +0000 UTC m=+933.722611512" watchObservedRunningTime="2025-12-08 15:01:52.624511556 +0000 UTC m=+933.724517671" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.628192 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.655480 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" event={"ID":"db804bba-5a84-4d23-93ca-e7118afc2af8","Type":"ContainerStarted","Data":"61425a5b37f577ab68371f6d433e07ca84f0c67a5fe3f62429de4869c9d9fad9"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.670294 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" event={"ID":"c91237dd-8bb9-4240-b24a-e19b58687d2e","Type":"ContainerStarted","Data":"7c08cbc9f21dba72f70a105f2eb8ac6b83a99e70d0ed5960b414a49a62cfbc37"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.689539 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-h2xw4" podStartSLOduration=5.02459033 podStartE2EDuration="44.689521094s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.63253801 +0000 UTC m=+891.732544125" lastFinishedPulling="2025-12-08 15:01:50.297468764 +0000 UTC m=+931.397474889" observedRunningTime="2025-12-08 15:01:52.687764589 +0000 UTC m=+933.787770704" watchObservedRunningTime="2025-12-08 15:01:52.689521094 +0000 UTC m=+933.789527209" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.702517 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" event={"ID":"5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8","Type":"ContainerStarted","Data":"47d254a6e402e8cb3b1c0d60b85f30618cd28869bfd9478fd141bc67c4ab31b8"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.703229 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.724955 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" event={"ID":"e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72","Type":"ContainerStarted","Data":"ffbfdc077c8657b8af4deda7eca535223f3e0a8c2b56f5ff554a4551fe53dab3"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.725989 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.744182 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.794041 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" event={"ID":"fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae","Type":"ContainerStarted","Data":"511461d4126d2f1038ca9173c0a11bcd68e3868d748d658dc17c500b9d147a8e"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.794879 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.799013 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" podStartSLOduration=12.552078843 podStartE2EDuration="44.798995252s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:11.091334858 +0000 UTC m=+892.191340973" lastFinishedPulling="2025-12-08 15:01:43.338251267 +0000 UTC m=+924.438257382" observedRunningTime="2025-12-08 15:01:52.741526991 +0000 UTC m=+933.841533106" watchObservedRunningTime="2025-12-08 15:01:52.798995252 +0000 UTC m=+933.899001367" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.807370 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" event={"ID":"8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76","Type":"ContainerStarted","Data":"5a73594e7a9e5f6d9f4318f09967c169379e7438409023d9179d7da91e01adda"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.807869 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.819195 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.819529 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-frz8t" podStartSLOduration=5.263290707 podStartE2EDuration="44.819512027s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.859170997 +0000 UTC m=+891.959177112" lastFinishedPulling="2025-12-08 15:01:50.415392317 +0000 UTC m=+931.515398432" observedRunningTime="2025-12-08 15:01:52.816137182 +0000 UTC m=+933.916143297" watchObservedRunningTime="2025-12-08 15:01:52.819512027 +0000 UTC m=+933.919518152" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.839099 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" event={"ID":"8b8eda01-4bc4-4eca-b909-901024ef466b","Type":"ContainerStarted","Data":"a5bc5ec9b9fd65ae22df51d24ba3f220c1cf7c27a31a832ca81838e95518901d"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.839880 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.887669 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" podStartSLOduration=38.758028982 podStartE2EDuration="44.887653823s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:44.069521626 +0000 UTC m=+925.169527741" lastFinishedPulling="2025-12-08 15:01:50.199146467 +0000 UTC m=+931.299152582" observedRunningTime="2025-12-08 15:01:52.885925739 +0000 UTC m=+933.985931854" watchObservedRunningTime="2025-12-08 15:01:52.887653823 +0000 UTC m=+933.987659928" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.895271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" event={"ID":"4ee4db87-8c97-4002-844b-6c55e1a41b2e","Type":"ContainerStarted","Data":"0e8047a658c25c19b3b49ac86cdd7cdc55167f3ae129d3214184ed4d006a1af3"} Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.896766 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.917295 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.939523 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-rnqzt" podStartSLOduration=5.553800674 podStartE2EDuration="44.939505456s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:11.078010818 +0000 UTC m=+892.178016933" lastFinishedPulling="2025-12-08 15:01:50.4637156 +0000 UTC m=+931.563721715" observedRunningTime="2025-12-08 15:01:52.912855147 +0000 UTC m=+934.012861262" watchObservedRunningTime="2025-12-08 15:01:52.939505456 +0000 UTC m=+934.039511571" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.998113 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" podStartSLOduration=12.699516906 podStartE2EDuration="44.998096282s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:11.043866064 +0000 UTC m=+892.143872179" lastFinishedPulling="2025-12-08 15:01:43.34244544 +0000 UTC m=+924.442451555" observedRunningTime="2025-12-08 15:01:52.972304329 +0000 UTC m=+934.072310444" watchObservedRunningTime="2025-12-08 15:01:52.998096282 +0000 UTC m=+934.098102397" Dec 08 15:01:52 crc kubenswrapper[4894]: I1208 15:01:52.999133 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-f67zc" podStartSLOduration=4.74418742 podStartE2EDuration="44.999122814s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.042506849 +0000 UTC m=+891.142512964" lastFinishedPulling="2025-12-08 15:01:50.297442243 +0000 UTC m=+931.397448358" observedRunningTime="2025-12-08 15:01:52.992245667 +0000 UTC m=+934.092251782" watchObservedRunningTime="2025-12-08 15:01:52.999122814 +0000 UTC m=+934.099128929" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.023044 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" podStartSLOduration=12.761454866 podStartE2EDuration="45.023027457s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:11.077851993 +0000 UTC m=+892.177858108" lastFinishedPulling="2025-12-08 15:01:43.339424584 +0000 UTC m=+924.439430699" observedRunningTime="2025-12-08 15:01:53.019982761 +0000 UTC m=+934.119988876" watchObservedRunningTime="2025-12-08 15:01:53.023027457 +0000 UTC m=+934.123033572" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.081916 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-8vb7c" podStartSLOduration=5.455769769 podStartE2EDuration="45.08188948s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.788482531 +0000 UTC m=+891.888488646" lastFinishedPulling="2025-12-08 15:01:50.414602242 +0000 UTC m=+931.514608357" observedRunningTime="2025-12-08 15:01:53.081264621 +0000 UTC m=+934.181270736" watchObservedRunningTime="2025-12-08 15:01:53.08188948 +0000 UTC m=+934.181895595" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.905936 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" event={"ID":"afda731d-73a8-443f-9647-a774ca8c3146","Type":"ContainerStarted","Data":"ed871e01adc53b9e1e5df37b60ac6ca51af9d075b32ccc11c82b281e70ae115d"} Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.907322 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.911186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" event={"ID":"2df93571-3b4e-4ab0-ad8b-34d2ef76034c","Type":"ContainerStarted","Data":"b982a2eb259f6f7c3773def6f0572024e2a450ad89f77182cbee3405066bcb19"} Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.911866 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.913534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" event={"ID":"28cfecb9-9c7c-49ae-af33-fc2d9acd374a","Type":"ContainerStarted","Data":"5042fbfc2f4c770ac373438aab21629398c67305a410a046af135ab59df5b56a"} Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.913960 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.914935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" event={"ID":"59051068-3252-47e6-a826-79d1951c3d5c","Type":"ContainerStarted","Data":"e080233d9e739a2fcdf121e993bdf949f741394568b1dc768cf5ea5f56eb9466"} Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.915100 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.916845 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" event={"ID":"f7ec7a43-f5b5-474f-af15-d0e664699477","Type":"ContainerStarted","Data":"d9a4f41393285c0a5e4c6a1f14ad3c318e93d6f62c3004dd2c179534666d2d85"} Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.917224 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.921170 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" event={"ID":"750c567a-693a-401b-8e7f-d8f1befbeacb","Type":"ContainerStarted","Data":"61700eea3f0605bff8ebed1f20f35ef658337d26cf36fae18cb6e98847c99845"} Dec 08 15:01:53 crc kubenswrapper[4894]: I1208 15:01:53.923355 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" podStartSLOduration=3.1804339329999998 podStartE2EDuration="45.923342279s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.352878752 +0000 UTC m=+891.452884877" lastFinishedPulling="2025-12-08 15:01:53.095787108 +0000 UTC m=+934.195793223" observedRunningTime="2025-12-08 15:01:53.922156542 +0000 UTC m=+935.022162647" watchObservedRunningTime="2025-12-08 15:01:53.923342279 +0000 UTC m=+935.023348394" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.052480 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" podStartSLOduration=5.198612729 podStartE2EDuration="46.052438834s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.959551648 +0000 UTC m=+892.059557763" lastFinishedPulling="2025-12-08 15:01:51.813377753 +0000 UTC m=+932.913383868" observedRunningTime="2025-12-08 15:01:54.047964344 +0000 UTC m=+935.147970459" watchObservedRunningTime="2025-12-08 15:01:54.052438834 +0000 UTC m=+935.152444949" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.072076 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" podStartSLOduration=4.947341777 podStartE2EDuration="46.072061572s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.788476581 +0000 UTC m=+891.888482696" lastFinishedPulling="2025-12-08 15:01:51.913196376 +0000 UTC m=+933.013202491" observedRunningTime="2025-12-08 15:01:54.066230399 +0000 UTC m=+935.166236514" watchObservedRunningTime="2025-12-08 15:01:54.072061572 +0000 UTC m=+935.172067687" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.086987 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" podStartSLOduration=4.973723278 podStartE2EDuration="46.086967642s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.788862073 +0000 UTC m=+891.888868188" lastFinishedPulling="2025-12-08 15:01:51.902106437 +0000 UTC m=+933.002112552" observedRunningTime="2025-12-08 15:01:54.084071491 +0000 UTC m=+935.184077606" watchObservedRunningTime="2025-12-08 15:01:54.086967642 +0000 UTC m=+935.186973757" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.108245 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" podStartSLOduration=2.635002658 podStartE2EDuration="46.108222271s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:09.621907314 +0000 UTC m=+890.721913429" lastFinishedPulling="2025-12-08 15:01:53.095126927 +0000 UTC m=+934.195133042" observedRunningTime="2025-12-08 15:01:54.107337683 +0000 UTC m=+935.207343798" watchObservedRunningTime="2025-12-08 15:01:54.108222271 +0000 UTC m=+935.208228386" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.127164 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" podStartSLOduration=4.674164874 podStartE2EDuration="46.127145167s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.356244819 +0000 UTC m=+891.456250934" lastFinishedPulling="2025-12-08 15:01:51.809225122 +0000 UTC m=+932.909231227" observedRunningTime="2025-12-08 15:01:54.123392489 +0000 UTC m=+935.223398604" watchObservedRunningTime="2025-12-08 15:01:54.127145167 +0000 UTC m=+935.227151282" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.929120 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" event={"ID":"db804bba-5a84-4d23-93ca-e7118afc2af8","Type":"ContainerStarted","Data":"bcd48b67ee3d181f8b8b2a2577ac18a270be205faadcd62efdfb37bab203ba86"} Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.929258 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.931628 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" event={"ID":"c91237dd-8bb9-4240-b24a-e19b58687d2e","Type":"ContainerStarted","Data":"95f37c27902bac19880605109611bba64b3a5c1f86c50972414e1ec30591ca3f"} Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.933268 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.957857 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" podStartSLOduration=4.618082067 podStartE2EDuration="46.957835097s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:11.042085357 +0000 UTC m=+892.142091472" lastFinishedPulling="2025-12-08 15:01:53.381838387 +0000 UTC m=+934.481844502" observedRunningTime="2025-12-08 15:01:54.950149554 +0000 UTC m=+936.050155679" watchObservedRunningTime="2025-12-08 15:01:54.957835097 +0000 UTC m=+936.057841222" Dec 08 15:01:54 crc kubenswrapper[4894]: I1208 15:01:54.969941 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" podStartSLOduration=4.502651742 podStartE2EDuration="46.969922667s" podCreationTimestamp="2025-12-08 15:01:08 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.911512765 +0000 UTC m=+892.011518880" lastFinishedPulling="2025-12-08 15:01:53.37878369 +0000 UTC m=+934.478789805" observedRunningTime="2025-12-08 15:01:54.968119341 +0000 UTC m=+936.068125456" watchObservedRunningTime="2025-12-08 15:01:54.969922667 +0000 UTC m=+936.069928792" Dec 08 15:01:55 crc kubenswrapper[4894]: I1208 15:01:55.077200 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879f4lcw4" Dec 08 15:01:55 crc kubenswrapper[4894]: I1208 15:01:55.955802 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" Dec 08 15:01:57 crc kubenswrapper[4894]: I1208 15:01:57.885269 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.015689 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" event={"ID":"9c6f4113-6d2f-4fce-b76b-8802b3b605c6","Type":"ContainerStarted","Data":"8748d6a9fe72d209a01fcd38289806081a00eb20f92317743cae8f0ac00d5452"} Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.026937 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.032829 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-97mkf" podStartSLOduration=3.057590738 podStartE2EDuration="49.032793832s" podCreationTimestamp="2025-12-08 15:01:09 +0000 UTC" firstStartedPulling="2025-12-08 15:01:10.858795745 +0000 UTC m=+891.958801870" lastFinishedPulling="2025-12-08 15:01:56.833998859 +0000 UTC m=+937.934004964" observedRunningTime="2025-12-08 15:01:58.029858029 +0000 UTC m=+939.129864144" watchObservedRunningTime="2025-12-08 15:01:58.032793832 +0000 UTC m=+939.132799957" Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.119517 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r6hn7"] Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.581437 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-kh5v9" Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.755694 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-mtkrb" Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.873620 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-47bpp" Dec 08 15:01:58 crc kubenswrapper[4894]: I1208 15:01:58.910954 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-w7xbj" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.022578 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r6hn7" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="registry-server" containerID="cri-o://20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649" gracePeriod=2 Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.113249 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-lbszt" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.271681 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2sfqd" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.313934 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-jc2rb" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.338978 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cqq4p" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.494545 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-sxfkp" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.516245 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-6qvtl" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.638361 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-69wkl" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.720784 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-zhwdn" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.916517 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.945212 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-catalog-content\") pod \"a377a91d-b683-4a28-9633-0c3f2367e3e5\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.945279 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-utilities\") pod \"a377a91d-b683-4a28-9633-0c3f2367e3e5\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.945332 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfp94\" (UniqueName: \"kubernetes.io/projected/a377a91d-b683-4a28-9633-0c3f2367e3e5-kube-api-access-sfp94\") pod \"a377a91d-b683-4a28-9633-0c3f2367e3e5\" (UID: \"a377a91d-b683-4a28-9633-0c3f2367e3e5\") " Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.947852 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-utilities" (OuterVolumeSpecName: "utilities") pod "a377a91d-b683-4a28-9633-0c3f2367e3e5" (UID: "a377a91d-b683-4a28-9633-0c3f2367e3e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:01:59 crc kubenswrapper[4894]: I1208 15:01:59.953726 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a377a91d-b683-4a28-9633-0c3f2367e3e5-kube-api-access-sfp94" (OuterVolumeSpecName: "kube-api-access-sfp94") pod "a377a91d-b683-4a28-9633-0c3f2367e3e5" (UID: "a377a91d-b683-4a28-9633-0c3f2367e3e5"). InnerVolumeSpecName "kube-api-access-sfp94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.033491 4894 generic.go:334] "Generic (PLEG): container finished" podID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerID="20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649" exitCode=0 Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.033559 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerDied","Data":"20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649"} Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.033578 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r6hn7" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.033626 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r6hn7" event={"ID":"a377a91d-b683-4a28-9633-0c3f2367e3e5","Type":"ContainerDied","Data":"bdfbdfb6ce441d1ad4fdee3f2f6cd28b5268ade8781d82a50a00f1d75c7bc441"} Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.033655 4894 scope.go:117] "RemoveContainer" containerID="20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.047308 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfp94\" (UniqueName: \"kubernetes.io/projected/a377a91d-b683-4a28-9633-0c3f2367e3e5-kube-api-access-sfp94\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.047346 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.055621 4894 scope.go:117] "RemoveContainer" containerID="a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.072314 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a377a91d-b683-4a28-9633-0c3f2367e3e5" (UID: "a377a91d-b683-4a28-9633-0c3f2367e3e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.075682 4894 scope.go:117] "RemoveContainer" containerID="cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.095289 4894 scope.go:117] "RemoveContainer" containerID="20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649" Dec 08 15:02:00 crc kubenswrapper[4894]: E1208 15:02:00.095731 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649\": container with ID starting with 20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649 not found: ID does not exist" containerID="20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.095778 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649"} err="failed to get container status \"20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649\": rpc error: code = NotFound desc = could not find container \"20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649\": container with ID starting with 20deac930e8f83a11f993e91c11d0eda4ccab4536d15e7a1bec7fb41221b3649 not found: ID does not exist" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.095805 4894 scope.go:117] "RemoveContainer" containerID="a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455" Dec 08 15:02:00 crc kubenswrapper[4894]: E1208 15:02:00.096166 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455\": container with ID starting with a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455 not found: ID does not exist" containerID="a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.096194 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455"} err="failed to get container status \"a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455\": rpc error: code = NotFound desc = could not find container \"a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455\": container with ID starting with a3a5fd30636051b416ca818ff50f9b5cb99efe082f8da9b6a17d63109f467455 not found: ID does not exist" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.096215 4894 scope.go:117] "RemoveContainer" containerID="cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d" Dec 08 15:02:00 crc kubenswrapper[4894]: E1208 15:02:00.096527 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d\": container with ID starting with cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d not found: ID does not exist" containerID="cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.096546 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d"} err="failed to get container status \"cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d\": rpc error: code = NotFound desc = could not find container \"cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d\": container with ID starting with cc2fc4a10d98fc0b6552aa51792d0cc4da500f0bd5b5afc0cfae74a24793e51d not found: ID does not exist" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.148717 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a377a91d-b683-4a28-9633-0c3f2367e3e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.362376 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r6hn7"] Dec 08 15:02:00 crc kubenswrapper[4894]: I1208 15:02:00.367643 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r6hn7"] Dec 08 15:02:01 crc kubenswrapper[4894]: I1208 15:02:01.207369 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" path="/var/lib/kubelet/pods/a377a91d-b683-4a28-9633-0c3f2367e3e5/volumes" Dec 08 15:02:01 crc kubenswrapper[4894]: I1208 15:02:01.745515 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8695b4b66-c64t7" Dec 08 15:02:04 crc kubenswrapper[4894]: I1208 15:02:04.449266 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-4xrts" Dec 08 15:02:07 crc kubenswrapper[4894]: I1208 15:02:07.297391 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:02:07 crc kubenswrapper[4894]: I1208 15:02:07.297717 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.043249 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8x6vs"] Dec 08 15:02:22 crc kubenswrapper[4894]: E1208 15:02:22.044142 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="registry-server" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.044163 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="registry-server" Dec 08 15:02:22 crc kubenswrapper[4894]: E1208 15:02:22.044205 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="extract-utilities" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.044217 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="extract-utilities" Dec 08 15:02:22 crc kubenswrapper[4894]: E1208 15:02:22.044248 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="extract-content" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.044258 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="extract-content" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.044445 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a377a91d-b683-4a28-9633-0c3f2367e3e5" containerName="registry-server" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.045651 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.047557 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.047801 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.048610 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.052060 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-khd8b" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.068060 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8x6vs"] Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.106349 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mg8lt"] Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.107629 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.109233 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.121255 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mg8lt"] Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.146004 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-kube-api-access-n8s5z\") pod \"dnsmasq-dns-675f4bcbfc-8x6vs\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.146071 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-config\") pod \"dnsmasq-dns-675f4bcbfc-8x6vs\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.247681 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.247973 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-kube-api-access-n8s5z\") pod \"dnsmasq-dns-675f4bcbfc-8x6vs\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.247996 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-config\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.248015 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ckrz\" (UniqueName: \"kubernetes.io/projected/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-kube-api-access-9ckrz\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.248040 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-config\") pod \"dnsmasq-dns-675f4bcbfc-8x6vs\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.248874 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-config\") pod \"dnsmasq-dns-675f4bcbfc-8x6vs\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.271181 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-kube-api-access-n8s5z\") pod \"dnsmasq-dns-675f4bcbfc-8x6vs\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.350251 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.350348 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-config\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.350381 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ckrz\" (UniqueName: \"kubernetes.io/projected/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-kube-api-access-9ckrz\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.351434 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-config\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.352679 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.364024 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.368669 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ckrz\" (UniqueName: \"kubernetes.io/projected/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-kube-api-access-9ckrz\") pod \"dnsmasq-dns-78dd6ddcc-mg8lt\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.422089 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.734309 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mg8lt"] Dec 08 15:02:22 crc kubenswrapper[4894]: I1208 15:02:22.814075 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8x6vs"] Dec 08 15:02:22 crc kubenswrapper[4894]: W1208 15:02:22.818289 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca81c74_cf4b_4685_99ef_8131fb54d6b7.slice/crio-cd296015553ecc262c6af7d9133e446aad7c882d194a7f5f91ba43954c6e93d4 WatchSource:0}: Error finding container cd296015553ecc262c6af7d9133e446aad7c882d194a7f5f91ba43954c6e93d4: Status 404 returned error can't find the container with id cd296015553ecc262c6af7d9133e446aad7c882d194a7f5f91ba43954c6e93d4 Dec 08 15:02:23 crc kubenswrapper[4894]: I1208 15:02:23.209045 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" event={"ID":"3ca81c74-cf4b-4685-99ef-8131fb54d6b7","Type":"ContainerStarted","Data":"cd296015553ecc262c6af7d9133e446aad7c882d194a7f5f91ba43954c6e93d4"} Dec 08 15:02:23 crc kubenswrapper[4894]: I1208 15:02:23.210112 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" event={"ID":"93eebb04-9aef-4e75-ad35-4d2c2c359f3e","Type":"ContainerStarted","Data":"a3076a7ac0d1acc7ecefd3ce3378fe457ad5a3824df1b25380d61543f8bdb5c3"} Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.233598 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8x6vs"] Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.259269 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j85bz"] Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.260690 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.273748 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j85bz"] Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.405099 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-config\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.405441 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs9m5\" (UniqueName: \"kubernetes.io/projected/7619d825-e819-4ab8-810f-584e402aa72c-kube-api-access-rs9m5\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.405503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.508164 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-config\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.508222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs9m5\" (UniqueName: \"kubernetes.io/projected/7619d825-e819-4ab8-810f-584e402aa72c-kube-api-access-rs9m5\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.508275 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.509138 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.509147 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-config\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.536238 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs9m5\" (UniqueName: \"kubernetes.io/projected/7619d825-e819-4ab8-810f-584e402aa72c-kube-api-access-rs9m5\") pod \"dnsmasq-dns-666b6646f7-j85bz\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.558724 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mg8lt"] Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.581467 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ffr6k"] Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.582596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.587671 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.602156 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ffr6k"] Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.710650 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-config\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.710703 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7kxn\" (UniqueName: \"kubernetes.io/projected/d164d26b-538c-4474-8c58-c840fecfdba6-kube-api-access-w7kxn\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.710744 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.811723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-config\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.811779 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7kxn\" (UniqueName: \"kubernetes.io/projected/d164d26b-538c-4474-8c58-c840fecfdba6-kube-api-access-w7kxn\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.811809 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.812640 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.812699 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-config\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.829141 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7kxn\" (UniqueName: \"kubernetes.io/projected/d164d26b-538c-4474-8c58-c840fecfdba6-kube-api-access-w7kxn\") pod \"dnsmasq-dns-57d769cc4f-ffr6k\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:25 crc kubenswrapper[4894]: I1208 15:02:25.899511 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.397609 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.398734 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.400785 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.400809 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.401253 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.401558 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.401687 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.401858 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.401929 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ttv88" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.419699 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520494 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520596 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d911c29-3d60-4990-b6ec-240d1acebc86-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520628 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d911c29-3d60-4990-b6ec-240d1acebc86-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520644 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520682 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520873 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520961 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.520982 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.521020 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsjbl\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-kube-api-access-tsjbl\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.521086 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622149 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622196 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622228 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsjbl\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-kube-api-access-tsjbl\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622301 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622315 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d911c29-3d60-4990-b6ec-240d1acebc86-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622340 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d911c29-3d60-4990-b6ec-240d1acebc86-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.622997 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.624136 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.624660 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.624889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.627564 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.627709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.636301 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d911c29-3d60-4990-b6ec-240d1acebc86-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.638249 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.650362 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.655080 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsjbl\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-kube-api-access-tsjbl\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.655963 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.683120 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d911c29-3d60-4990-b6ec-240d1acebc86-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.698459 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.700017 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.709283 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.709504 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.709804 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.709950 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.710045 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.710117 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-544nh" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.711052 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.711290 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.726661 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835437 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835502 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835530 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbskj\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-kube-api-access-fbskj\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835587 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835661 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835681 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835735 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835824 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.835853 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938443 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938679 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938722 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938748 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938765 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938803 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938849 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938910 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.938997 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.939023 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbskj\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-kube-api-access-fbskj\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.940277 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.943441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.946535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.947865 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.948679 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.948796 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.952126 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.954509 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.965304 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.968465 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbskj\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-kube-api-access-fbskj\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:26 crc kubenswrapper[4894]: I1208 15:02:26.969725 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:27 crc kubenswrapper[4894]: I1208 15:02:27.010246 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:27 crc kubenswrapper[4894]: I1208 15:02:27.056216 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.388433 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.390488 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.393033 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-k5njr" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.393557 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.394708 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.395618 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.411380 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.417399 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470428 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755a6729-c028-4206-abc7-5b4bd3825a3e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470510 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cnbk\" (UniqueName: \"kubernetes.io/projected/755a6729-c028-4206-abc7-5b4bd3825a3e-kube-api-access-4cnbk\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470534 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/755a6729-c028-4206-abc7-5b4bd3825a3e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470553 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/755a6729-c028-4206-abc7-5b4bd3825a3e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470578 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-kolla-config\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470753 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-config-data-default\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.470838 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577683 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-config-data-default\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577741 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577901 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755a6729-c028-4206-abc7-5b4bd3825a3e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577919 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cnbk\" (UniqueName: \"kubernetes.io/projected/755a6729-c028-4206-abc7-5b4bd3825a3e-kube-api-access-4cnbk\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/755a6729-c028-4206-abc7-5b4bd3825a3e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577957 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/755a6729-c028-4206-abc7-5b4bd3825a3e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.577982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-kolla-config\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.578291 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.578884 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-kolla-config\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.578956 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-config-data-default\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.579159 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/755a6729-c028-4206-abc7-5b4bd3825a3e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.579680 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/755a6729-c028-4206-abc7-5b4bd3825a3e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.584523 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/755a6729-c028-4206-abc7-5b4bd3825a3e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.584658 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755a6729-c028-4206-abc7-5b4bd3825a3e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.611380 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cnbk\" (UniqueName: \"kubernetes.io/projected/755a6729-c028-4206-abc7-5b4bd3825a3e-kube-api-access-4cnbk\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.715880 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"755a6729-c028-4206-abc7-5b4bd3825a3e\") " pod="openstack/openstack-galera-0" Dec 08 15:02:28 crc kubenswrapper[4894]: I1208 15:02:28.723143 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.670025 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.671533 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.673483 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.673628 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-czhkv" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.673836 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.673961 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.683266 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702624 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702670 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e121dc9-bd10-476c-b28b-06c6dcb09165-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702704 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702735 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702757 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e121dc9-bd10-476c-b28b-06c6dcb09165-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702789 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702805 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1e121dc9-bd10-476c-b28b-06c6dcb09165-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.702841 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thkgk\" (UniqueName: \"kubernetes.io/projected/1e121dc9-bd10-476c-b28b-06c6dcb09165-kube-api-access-thkgk\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804288 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804342 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1e121dc9-bd10-476c-b28b-06c6dcb09165-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804372 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thkgk\" (UniqueName: \"kubernetes.io/projected/1e121dc9-bd10-476c-b28b-06c6dcb09165-kube-api-access-thkgk\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804438 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e121dc9-bd10-476c-b28b-06c6dcb09165-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804506 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804535 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804562 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e121dc9-bd10-476c-b28b-06c6dcb09165-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.804849 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1e121dc9-bd10-476c-b28b-06c6dcb09165-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.805223 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.806179 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.806366 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.806618 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e121dc9-bd10-476c-b28b-06c6dcb09165-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.820044 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e121dc9-bd10-476c-b28b-06c6dcb09165-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.820655 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e121dc9-bd10-476c-b28b-06c6dcb09165-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.854712 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thkgk\" (UniqueName: \"kubernetes.io/projected/1e121dc9-bd10-476c-b28b-06c6dcb09165-kube-api-access-thkgk\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.863195 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1e121dc9-bd10-476c-b28b-06c6dcb09165\") " pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.936466 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.937982 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.947126 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nmlkj" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.947337 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.948025 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 08 15:02:29 crc kubenswrapper[4894]: I1208 15:02:29.966841 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.003540 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.007612 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn7bw\" (UniqueName: \"kubernetes.io/projected/637635c4-70df-45dd-8139-9c8dc3f53ea0-kube-api-access-xn7bw\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.007679 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/637635c4-70df-45dd-8139-9c8dc3f53ea0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.007703 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/637635c4-70df-45dd-8139-9c8dc3f53ea0-config-data\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.007896 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/637635c4-70df-45dd-8139-9c8dc3f53ea0-kolla-config\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.007971 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/637635c4-70df-45dd-8139-9c8dc3f53ea0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.108841 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn7bw\" (UniqueName: \"kubernetes.io/projected/637635c4-70df-45dd-8139-9c8dc3f53ea0-kube-api-access-xn7bw\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.108927 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/637635c4-70df-45dd-8139-9c8dc3f53ea0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.108954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/637635c4-70df-45dd-8139-9c8dc3f53ea0-config-data\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.108989 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/637635c4-70df-45dd-8139-9c8dc3f53ea0-kolla-config\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.109012 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/637635c4-70df-45dd-8139-9c8dc3f53ea0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.110227 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/637635c4-70df-45dd-8139-9c8dc3f53ea0-config-data\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.110279 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/637635c4-70df-45dd-8139-9c8dc3f53ea0-kolla-config\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.113317 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/637635c4-70df-45dd-8139-9c8dc3f53ea0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.117494 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/637635c4-70df-45dd-8139-9c8dc3f53ea0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.132446 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn7bw\" (UniqueName: \"kubernetes.io/projected/637635c4-70df-45dd-8139-9c8dc3f53ea0-kube-api-access-xn7bw\") pod \"memcached-0\" (UID: \"637635c4-70df-45dd-8139-9c8dc3f53ea0\") " pod="openstack/memcached-0" Dec 08 15:02:30 crc kubenswrapper[4894]: I1208 15:02:30.280239 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.216752 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.774510 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.775689 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.778198 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-9cs5c" Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.781471 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.845674 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjnzc\" (UniqueName: \"kubernetes.io/projected/102cc0a0-d75c-4050-a100-6c97351084dd-kube-api-access-hjnzc\") pod \"kube-state-metrics-0\" (UID: \"102cc0a0-d75c-4050-a100-6c97351084dd\") " pod="openstack/kube-state-metrics-0" Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.947189 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjnzc\" (UniqueName: \"kubernetes.io/projected/102cc0a0-d75c-4050-a100-6c97351084dd-kube-api-access-hjnzc\") pod \"kube-state-metrics-0\" (UID: \"102cc0a0-d75c-4050-a100-6c97351084dd\") " pod="openstack/kube-state-metrics-0" Dec 08 15:02:31 crc kubenswrapper[4894]: I1208 15:02:31.965678 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjnzc\" (UniqueName: \"kubernetes.io/projected/102cc0a0-d75c-4050-a100-6c97351084dd-kube-api-access-hjnzc\") pod \"kube-state-metrics-0\" (UID: \"102cc0a0-d75c-4050-a100-6c97351084dd\") " pod="openstack/kube-state-metrics-0" Dec 08 15:02:32 crc kubenswrapper[4894]: I1208 15:02:32.107224 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 15:02:33 crc kubenswrapper[4894]: W1208 15:02:33.808335 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4dd5876_f3d0_4256_9524_ff3bf39ea28a.slice/crio-596f1a6693a6a0e395b0773bfbc593ee4ff2afd5ae80e7565c7d0b18d7940f2d WatchSource:0}: Error finding container 596f1a6693a6a0e395b0773bfbc593ee4ff2afd5ae80e7565c7d0b18d7940f2d: Status 404 returned error can't find the container with id 596f1a6693a6a0e395b0773bfbc593ee4ff2afd5ae80e7565c7d0b18d7940f2d Dec 08 15:02:34 crc kubenswrapper[4894]: I1208 15:02:34.292888 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a4dd5876-f3d0-4256-9524-ff3bf39ea28a","Type":"ContainerStarted","Data":"596f1a6693a6a0e395b0773bfbc593ee4ff2afd5ae80e7565c7d0b18d7940f2d"} Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.253392 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-89p8p"] Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.255137 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.257752 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.259177 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-hmwd6" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.259390 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.266376 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-89p8p"] Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.302685 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-run\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.305734 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkkq9\" (UniqueName: \"kubernetes.io/projected/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-kube-api-access-nkkq9\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.305791 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-scripts\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.305882 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-run-ovn\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.306038 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-ovn-controller-tls-certs\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.306128 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-log-ovn\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.306154 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-combined-ca-bundle\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.320081 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-zxxl4"] Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.325789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.336230 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zxxl4"] Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-log-ovn\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407832 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-combined-ca-bundle\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407864 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-run\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407887 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-run\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407916 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1996355-2ff3-49de-b526-a5cef3dfc253-scripts\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407939 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkkq9\" (UniqueName: \"kubernetes.io/projected/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-kube-api-access-nkkq9\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407956 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-scripts\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.407983 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-run-ovn\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.408003 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-log\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.408029 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s99b\" (UniqueName: \"kubernetes.io/projected/e1996355-2ff3-49de-b526-a5cef3dfc253-kube-api-access-5s99b\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.408046 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-lib\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.408070 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-etc-ovs\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.408095 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-ovn-controller-tls-certs\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.409503 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-log-ovn\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.409694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-run\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.411653 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-scripts\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.414617 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-var-run-ovn\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.418568 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-ovn-controller-tls-certs\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.423680 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-combined-ca-bundle\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.424127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkkq9\" (UniqueName: \"kubernetes.io/projected/9d8ef054-d4aa-482b-abad-c961a2a5f2ad-kube-api-access-nkkq9\") pod \"ovn-controller-89p8p\" (UID: \"9d8ef054-d4aa-482b-abad-c961a2a5f2ad\") " pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.509672 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-log\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.510364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s99b\" (UniqueName: \"kubernetes.io/projected/e1996355-2ff3-49de-b526-a5cef3dfc253-kube-api-access-5s99b\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.510393 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-lib\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.510422 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-etc-ovs\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.510477 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-run\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.510507 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1996355-2ff3-49de-b526-a5cef3dfc253-scripts\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.510314 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-log\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.511472 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-etc-ovs\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.511526 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-run\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.511730 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1996355-2ff3-49de-b526-a5cef3dfc253-var-lib\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.513616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1996355-2ff3-49de-b526-a5cef3dfc253-scripts\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.541538 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s99b\" (UniqueName: \"kubernetes.io/projected/e1996355-2ff3-49de-b526-a5cef3dfc253-kube-api-access-5s99b\") pod \"ovn-controller-ovs-zxxl4\" (UID: \"e1996355-2ff3-49de-b526-a5cef3dfc253\") " pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.584758 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-89p8p" Dec 08 15:02:35 crc kubenswrapper[4894]: I1208 15:02:35.644786 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.124420 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.126301 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.130071 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.130284 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.130989 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.131275 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.132585 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-qrxgp" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.140985 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.223171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.223223 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.223274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.223358 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0949aa02-5e23-4bd9-9a93-180a59f701ad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.223391 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0949aa02-5e23-4bd9-9a93-180a59f701ad-config\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.223432 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0949aa02-5e23-4bd9-9a93-180a59f701ad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.224288 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.224360 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxk8s\" (UniqueName: \"kubernetes.io/projected/0949aa02-5e23-4bd9-9a93-180a59f701ad-kube-api-access-bxk8s\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326536 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0949aa02-5e23-4bd9-9a93-180a59f701ad-config\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326593 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0949aa02-5e23-4bd9-9a93-180a59f701ad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326620 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326646 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxk8s\" (UniqueName: \"kubernetes.io/projected/0949aa02-5e23-4bd9-9a93-180a59f701ad-kube-api-access-bxk8s\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326730 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326759 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326790 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.326892 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0949aa02-5e23-4bd9-9a93-180a59f701ad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.327510 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0949aa02-5e23-4bd9-9a93-180a59f701ad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.332509 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0949aa02-5e23-4bd9-9a93-180a59f701ad-config\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.333518 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.335084 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0949aa02-5e23-4bd9-9a93-180a59f701ad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.339716 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.340010 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.340760 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0949aa02-5e23-4bd9-9a93-180a59f701ad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.349217 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxk8s\" (UniqueName: \"kubernetes.io/projected/0949aa02-5e23-4bd9-9a93-180a59f701ad-kube-api-access-bxk8s\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.357359 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"0949aa02-5e23-4bd9-9a93-180a59f701ad\") " pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:36 crc kubenswrapper[4894]: I1208 15:02:36.445300 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 08 15:02:37 crc kubenswrapper[4894]: I1208 15:02:37.296682 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:02:37 crc kubenswrapper[4894]: I1208 15:02:37.297337 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:02:37 crc kubenswrapper[4894]: I1208 15:02:37.297452 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:02:37 crc kubenswrapper[4894]: I1208 15:02:37.298197 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b3439967813e9136ef0135406ae68191aab690f31fec7c75b03962a836112d1"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:02:37 crc kubenswrapper[4894]: I1208 15:02:37.298314 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://1b3439967813e9136ef0135406ae68191aab690f31fec7c75b03962a836112d1" gracePeriod=600 Dec 08 15:02:38 crc kubenswrapper[4894]: I1208 15:02:38.444406 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="1b3439967813e9136ef0135406ae68191aab690f31fec7c75b03962a836112d1" exitCode=0 Dec 08 15:02:38 crc kubenswrapper[4894]: I1208 15:02:38.444462 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"1b3439967813e9136ef0135406ae68191aab690f31fec7c75b03962a836112d1"} Dec 08 15:02:38 crc kubenswrapper[4894]: I1208 15:02:38.444509 4894 scope.go:117] "RemoveContainer" containerID="dfc08546d7b8c2b8b1aa3ec04284ba529dbbd4025e73389fcf8cb4e001930019" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.675193 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.680125 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.685633 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-ch9bf" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.686862 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.686943 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.686952 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.692201 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31bac65c-93e6-4816-837d-a0f467bb3966-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870472 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31bac65c-93e6-4816-837d-a0f467bb3966-config\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870696 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870829 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2crl\" (UniqueName: \"kubernetes.io/projected/31bac65c-93e6-4816-837d-a0f467bb3966-kube-api-access-z2crl\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870861 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/31bac65c-93e6-4816-837d-a0f467bb3966-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870897 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.870948 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.973550 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31bac65c-93e6-4816-837d-a0f467bb3966-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.973608 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.974225 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31bac65c-93e6-4816-837d-a0f467bb3966-config\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.974300 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.974392 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2crl\" (UniqueName: \"kubernetes.io/projected/31bac65c-93e6-4816-837d-a0f467bb3966-kube-api-access-z2crl\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.974416 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/31bac65c-93e6-4816-837d-a0f467bb3966-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.974558 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.974901 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/31bac65c-93e6-4816-837d-a0f467bb3966-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.975232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31bac65c-93e6-4816-837d-a0f467bb3966-config\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.975522 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31bac65c-93e6-4816-837d-a0f467bb3966-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.975851 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.976928 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.981165 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.981787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.986854 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/31bac65c-93e6-4816-837d-a0f467bb3966-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:39 crc kubenswrapper[4894]: I1208 15:02:39.997223 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2crl\" (UniqueName: \"kubernetes.io/projected/31bac65c-93e6-4816-837d-a0f467bb3966-kube-api-access-z2crl\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:40 crc kubenswrapper[4894]: I1208 15:02:40.015110 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"31bac65c-93e6-4816-837d-a0f467bb3966\") " pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:40 crc kubenswrapper[4894]: I1208 15:02:40.306978 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 08 15:02:41 crc kubenswrapper[4894]: E1208 15:02:41.654016 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 08 15:02:41 crc kubenswrapper[4894]: E1208 15:02:41.659008 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n8s5z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-8x6vs_openstack(3ca81c74-cf4b-4685-99ef-8131fb54d6b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:02:41 crc kubenswrapper[4894]: E1208 15:02:41.660589 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" podUID="3ca81c74-cf4b-4685-99ef-8131fb54d6b7" Dec 08 15:02:41 crc kubenswrapper[4894]: E1208 15:02:41.673929 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 08 15:02:41 crc kubenswrapper[4894]: E1208 15:02:41.674167 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9ckrz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-mg8lt_openstack(93eebb04-9aef-4e75-ad35-4d2c2c359f3e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:02:41 crc kubenswrapper[4894]: E1208 15:02:41.675268 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" podUID="93eebb04-9aef-4e75-ad35-4d2c2c359f3e" Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.037882 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.589633 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.598128 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j85bz"] Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.612497 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 08 15:02:42 crc kubenswrapper[4894]: W1208 15:02:42.615056 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod755a6729_c028_4206_abc7_5b4bd3825a3e.slice/crio-478cdfca9d73aebfe306ea58a0a7876ba6f8378ea6f0fadd07c2d029b2385c36 WatchSource:0}: Error finding container 478cdfca9d73aebfe306ea58a0a7876ba6f8378ea6f0fadd07c2d029b2385c36: Status 404 returned error can't find the container with id 478cdfca9d73aebfe306ea58a0a7876ba6f8378ea6f0fadd07c2d029b2385c36 Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.626586 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"9002401f278cd2ddc27524b106a2357f8f007e534b8798b7fd2145f577807600"} Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.630590 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d911c29-3d60-4990-b6ec-240d1acebc86","Type":"ContainerStarted","Data":"5f91cd22f57fc82e1033fe01b97cae861e9b914df86c4a426df19956ed94501e"} Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.633655 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"637635c4-70df-45dd-8139-9c8dc3f53ea0","Type":"ContainerStarted","Data":"8a5734f9bfd8ecdba0348a60655fef105e0063631d18e59cb0dbb6b7268c92e6"} Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.727169 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zxxl4"] Dec 08 15:02:42 crc kubenswrapper[4894]: W1208 15:02:42.739003 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1996355_2ff3_49de_b526_a5cef3dfc253.slice/crio-a58dfd70db6086a909d097e7a441d7d1403b3a71003685c51a5c2516046e1437 WatchSource:0}: Error finding container a58dfd70db6086a909d097e7a441d7d1403b3a71003685c51a5c2516046e1437: Status 404 returned error can't find the container with id a58dfd70db6086a909d097e7a441d7d1403b3a71003685c51a5c2516046e1437 Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.985623 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:02:42 crc kubenswrapper[4894]: I1208 15:02:42.994420 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.001423 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ffr6k"] Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.007250 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-89p8p"] Dec 08 15:02:43 crc kubenswrapper[4894]: W1208 15:02:43.008442 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e121dc9_bd10_476c_b28b_06c6dcb09165.slice/crio-b2b3442412701b88efef982fbb61ac5b89554f74e9f51b090ca7bc5631fa4bc4 WatchSource:0}: Error finding container b2b3442412701b88efef982fbb61ac5b89554f74e9f51b090ca7bc5631fa4bc4: Status 404 returned error can't find the container with id b2b3442412701b88efef982fbb61ac5b89554f74e9f51b090ca7bc5631fa4bc4 Dec 08 15:02:43 crc kubenswrapper[4894]: W1208 15:02:43.023674 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod102cc0a0_d75c_4050_a100_6c97351084dd.slice/crio-b5c06232fa14fdfb018353f51c59e8ebf8256a7ed50e052b8cdfe2c1291ab716 WatchSource:0}: Error finding container b5c06232fa14fdfb018353f51c59e8ebf8256a7ed50e052b8cdfe2c1291ab716: Status 404 returned error can't find the container with id b5c06232fa14fdfb018353f51c59e8ebf8256a7ed50e052b8cdfe2c1291ab716 Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.173121 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.178490 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.328224 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 08 15:02:43 crc kubenswrapper[4894]: W1208 15:02:43.341991 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0949aa02_5e23_4bd9_9a93_180a59f701ad.slice/crio-00fc49d361baca51d9b2d729586b58374fa106a4a7b07626857e3bd0c8a1050e WatchSource:0}: Error finding container 00fc49d361baca51d9b2d729586b58374fa106a4a7b07626857e3bd0c8a1050e: Status 404 returned error can't find the container with id 00fc49d361baca51d9b2d729586b58374fa106a4a7b07626857e3bd0c8a1050e Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.342615 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-config\") pod \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.343028 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ckrz\" (UniqueName: \"kubernetes.io/projected/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-kube-api-access-9ckrz\") pod \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.343065 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-kube-api-access-n8s5z\") pod \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\" (UID: \"3ca81c74-cf4b-4685-99ef-8131fb54d6b7\") " Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.343129 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-config" (OuterVolumeSpecName: "config") pod "3ca81c74-cf4b-4685-99ef-8131fb54d6b7" (UID: "3ca81c74-cf4b-4685-99ef-8131fb54d6b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.343146 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-dns-svc\") pod \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.343168 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-config\") pod \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\" (UID: \"93eebb04-9aef-4e75-ad35-4d2c2c359f3e\") " Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.343921 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.344205 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93eebb04-9aef-4e75-ad35-4d2c2c359f3e" (UID: "93eebb04-9aef-4e75-ad35-4d2c2c359f3e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.344598 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-config" (OuterVolumeSpecName: "config") pod "93eebb04-9aef-4e75-ad35-4d2c2c359f3e" (UID: "93eebb04-9aef-4e75-ad35-4d2c2c359f3e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.347046 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-kube-api-access-n8s5z" (OuterVolumeSpecName: "kube-api-access-n8s5z") pod "3ca81c74-cf4b-4685-99ef-8131fb54d6b7" (UID: "3ca81c74-cf4b-4685-99ef-8131fb54d6b7"). InnerVolumeSpecName "kube-api-access-n8s5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.348211 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-kube-api-access-9ckrz" (OuterVolumeSpecName: "kube-api-access-9ckrz") pod "93eebb04-9aef-4e75-ad35-4d2c2c359f3e" (UID: "93eebb04-9aef-4e75-ad35-4d2c2c359f3e"). InnerVolumeSpecName "kube-api-access-9ckrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.445949 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ckrz\" (UniqueName: \"kubernetes.io/projected/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-kube-api-access-9ckrz\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.446603 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8s5z\" (UniqueName: \"kubernetes.io/projected/3ca81c74-cf4b-4685-99ef-8131fb54d6b7-kube-api-access-n8s5z\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.446613 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.446625 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93eebb04-9aef-4e75-ad35-4d2c2c359f3e-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.643571 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" event={"ID":"3ca81c74-cf4b-4685-99ef-8131fb54d6b7","Type":"ContainerDied","Data":"cd296015553ecc262c6af7d9133e446aad7c882d194a7f5f91ba43954c6e93d4"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.643638 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8x6vs" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.646671 4894 generic.go:334] "Generic (PLEG): container finished" podID="7619d825-e819-4ab8-810f-584e402aa72c" containerID="2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5" exitCode=0 Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.646738 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" event={"ID":"7619d825-e819-4ab8-810f-584e402aa72c","Type":"ContainerDied","Data":"2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.646759 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" event={"ID":"7619d825-e819-4ab8-810f-584e402aa72c","Type":"ContainerStarted","Data":"82452a7eb2bbb8fbdc941229f93949c7de811f9802b48e6e90bccd8a11a6226a"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.650077 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zxxl4" event={"ID":"e1996355-2ff3-49de-b526-a5cef3dfc253","Type":"ContainerStarted","Data":"a58dfd70db6086a909d097e7a441d7d1403b3a71003685c51a5c2516046e1437"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.652040 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" event={"ID":"93eebb04-9aef-4e75-ad35-4d2c2c359f3e","Type":"ContainerDied","Data":"a3076a7ac0d1acc7ecefd3ce3378fe457ad5a3824df1b25380d61543f8bdb5c3"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.652113 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mg8lt" Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.659228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-89p8p" event={"ID":"9d8ef054-d4aa-482b-abad-c961a2a5f2ad","Type":"ContainerStarted","Data":"9fd6a5de90e447c4575300da518da4a30f247f826cc69db692823958c7a35f8e"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.660363 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"102cc0a0-d75c-4050-a100-6c97351084dd","Type":"ContainerStarted","Data":"b5c06232fa14fdfb018353f51c59e8ebf8256a7ed50e052b8cdfe2c1291ab716"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.664571 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1e121dc9-bd10-476c-b28b-06c6dcb09165","Type":"ContainerStarted","Data":"b2b3442412701b88efef982fbb61ac5b89554f74e9f51b090ca7bc5631fa4bc4"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.667932 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" event={"ID":"d164d26b-538c-4474-8c58-c840fecfdba6","Type":"ContainerStarted","Data":"6e4c16cdfc113be793315270553370a7e023973a3c0a16ae542ed065d140d1d2"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.671304 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"755a6729-c028-4206-abc7-5b4bd3825a3e","Type":"ContainerStarted","Data":"478cdfca9d73aebfe306ea58a0a7876ba6f8378ea6f0fadd07c2d029b2385c36"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.673091 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0949aa02-5e23-4bd9-9a93-180a59f701ad","Type":"ContainerStarted","Data":"00fc49d361baca51d9b2d729586b58374fa106a4a7b07626857e3bd0c8a1050e"} Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.721886 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8x6vs"] Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.735894 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8x6vs"] Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.766679 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mg8lt"] Dec 08 15:02:43 crc kubenswrapper[4894]: I1208 15:02:43.777547 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mg8lt"] Dec 08 15:02:44 crc kubenswrapper[4894]: I1208 15:02:44.093857 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 08 15:02:45 crc kubenswrapper[4894]: I1208 15:02:45.206647 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca81c74-cf4b-4685-99ef-8131fb54d6b7" path="/var/lib/kubelet/pods/3ca81c74-cf4b-4685-99ef-8131fb54d6b7/volumes" Dec 08 15:02:45 crc kubenswrapper[4894]: I1208 15:02:45.207039 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93eebb04-9aef-4e75-ad35-4d2c2c359f3e" path="/var/lib/kubelet/pods/93eebb04-9aef-4e75-ad35-4d2c2c359f3e/volumes" Dec 08 15:02:45 crc kubenswrapper[4894]: I1208 15:02:45.694649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"31bac65c-93e6-4816-837d-a0f467bb3966","Type":"ContainerStarted","Data":"fc41e8477b1e5fdceda1fa2bc5c5720669568652eea8eb3fcec307daf7b2c211"} Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.741216 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"637635c4-70df-45dd-8139-9c8dc3f53ea0","Type":"ContainerStarted","Data":"6e467af1d0243d02a0282c8562b1a8c4118c0c6d9c5a00843ea6de89a9f550df"} Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.741969 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.742769 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1e121dc9-bd10-476c-b28b-06c6dcb09165","Type":"ContainerStarted","Data":"6c23e988b134ff3e80b86cc69f41fa9833cd45e15bf6806c30a6e96245fbd1cf"} Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.744697 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"755a6729-c028-4206-abc7-5b4bd3825a3e","Type":"ContainerStarted","Data":"86c7359c65ade4a8c61efcae3cdf8e65cacdc9569d2321e180789ea399d4c312"} Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.759857 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" event={"ID":"7619d825-e819-4ab8-810f-584e402aa72c","Type":"ContainerStarted","Data":"de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593"} Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.761119 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.766065 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.546739656 podStartE2EDuration="21.766045342s" podCreationTimestamp="2025-12-08 15:02:29 +0000 UTC" firstStartedPulling="2025-12-08 15:02:42.603603007 +0000 UTC m=+983.703609132" lastFinishedPulling="2025-12-08 15:02:49.822908703 +0000 UTC m=+990.922914818" observedRunningTime="2025-12-08 15:02:50.758640199 +0000 UTC m=+991.858646314" watchObservedRunningTime="2025-12-08 15:02:50.766045342 +0000 UTC m=+991.866051457" Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.769892 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zxxl4" event={"ID":"e1996355-2ff3-49de-b526-a5cef3dfc253","Type":"ContainerStarted","Data":"c6b2e7a4d950b4bac37dab9156f3d9db21b3934970a196227ec20cd930a90578"} Dec 08 15:02:50 crc kubenswrapper[4894]: I1208 15:02:50.862393 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" podStartSLOduration=25.354140459 podStartE2EDuration="25.862371911s" podCreationTimestamp="2025-12-08 15:02:25 +0000 UTC" firstStartedPulling="2025-12-08 15:02:42.6202416 +0000 UTC m=+983.720247715" lastFinishedPulling="2025-12-08 15:02:43.128473052 +0000 UTC m=+984.228479167" observedRunningTime="2025-12-08 15:02:50.858599592 +0000 UTC m=+991.958605717" watchObservedRunningTime="2025-12-08 15:02:50.862371911 +0000 UTC m=+991.962378026" Dec 08 15:02:51 crc kubenswrapper[4894]: I1208 15:02:51.778941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0949aa02-5e23-4bd9-9a93-180a59f701ad","Type":"ContainerStarted","Data":"0dd89016bb5493a7d187c199728fe6dfe05d9fe1f7d8500f1524b9c1cbbd4e40"} Dec 08 15:02:52 crc kubenswrapper[4894]: E1208 15:02:52.489783 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd164d26b_538c_4474_8c58_c840fecfdba6.slice/crio-conmon-6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a.scope\": RecentStats: unable to find data in memory cache]" Dec 08 15:02:52 crc kubenswrapper[4894]: I1208 15:02:52.786400 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d911c29-3d60-4990-b6ec-240d1acebc86","Type":"ContainerStarted","Data":"340018ff56d632188be324a9d2eb44cf1163ca716d83f89365cb7acedd19ccc4"} Dec 08 15:02:52 crc kubenswrapper[4894]: I1208 15:02:52.787765 4894 generic.go:334] "Generic (PLEG): container finished" podID="e1996355-2ff3-49de-b526-a5cef3dfc253" containerID="c6b2e7a4d950b4bac37dab9156f3d9db21b3934970a196227ec20cd930a90578" exitCode=0 Dec 08 15:02:52 crc kubenswrapper[4894]: I1208 15:02:52.787807 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zxxl4" event={"ID":"e1996355-2ff3-49de-b526-a5cef3dfc253","Type":"ContainerDied","Data":"c6b2e7a4d950b4bac37dab9156f3d9db21b3934970a196227ec20cd930a90578"} Dec 08 15:02:52 crc kubenswrapper[4894]: I1208 15:02:52.791065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a4dd5876-f3d0-4256-9524-ff3bf39ea28a","Type":"ContainerStarted","Data":"8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d"} Dec 08 15:02:52 crc kubenswrapper[4894]: I1208 15:02:52.792499 4894 generic.go:334] "Generic (PLEG): container finished" podID="d164d26b-538c-4474-8c58-c840fecfdba6" containerID="6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a" exitCode=0 Dec 08 15:02:52 crc kubenswrapper[4894]: I1208 15:02:52.792579 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" event={"ID":"d164d26b-538c-4474-8c58-c840fecfdba6","Type":"ContainerDied","Data":"6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a"} Dec 08 15:02:55 crc kubenswrapper[4894]: I1208 15:02:55.281491 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 08 15:02:55 crc kubenswrapper[4894]: I1208 15:02:55.592033 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:02:55 crc kubenswrapper[4894]: I1208 15:02:55.815096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"31bac65c-93e6-4816-837d-a0f467bb3966","Type":"ContainerStarted","Data":"910510a8359f123b84e0b802f0ccd69c22ee8bf85d3d4eadcb3d48878fb5f912"} Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.824571 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-89p8p" event={"ID":"9d8ef054-d4aa-482b-abad-c961a2a5f2ad","Type":"ContainerStarted","Data":"65a3e223fd626d9524c491fe2feff40a2ce534bcb659aaf0abf4e76ff3995dd2"} Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.824982 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-89p8p" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.826236 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"102cc0a0-d75c-4050-a100-6c97351084dd","Type":"ContainerStarted","Data":"2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff"} Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.827010 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.830844 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" event={"ID":"d164d26b-538c-4474-8c58-c840fecfdba6","Type":"ContainerStarted","Data":"3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542"} Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.830983 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.833492 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zxxl4" event={"ID":"e1996355-2ff3-49de-b526-a5cef3dfc253","Type":"ContainerStarted","Data":"4ff8780d383b735d741da03f4d7c6589e3c958612921c10b679357a98d68e829"} Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.833530 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zxxl4" event={"ID":"e1996355-2ff3-49de-b526-a5cef3dfc253","Type":"ContainerStarted","Data":"c628a596d15db0a4a3c20d7a2c7c7b0f976bc62ea76dcbd22e574b79a9515b1b"} Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.833994 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.834043 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.854515 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-89p8p" podStartSLOduration=15.011986699 podStartE2EDuration="21.854491236s" podCreationTimestamp="2025-12-08 15:02:35 +0000 UTC" firstStartedPulling="2025-12-08 15:02:43.053330069 +0000 UTC m=+984.153336184" lastFinishedPulling="2025-12-08 15:02:49.895834606 +0000 UTC m=+990.995840721" observedRunningTime="2025-12-08 15:02:56.844408579 +0000 UTC m=+997.944414704" watchObservedRunningTime="2025-12-08 15:02:56.854491236 +0000 UTC m=+997.954497351" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.874247 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-zxxl4" podStartSLOduration=14.821926113 podStartE2EDuration="21.874229827s" podCreationTimestamp="2025-12-08 15:02:35 +0000 UTC" firstStartedPulling="2025-12-08 15:02:42.755621187 +0000 UTC m=+983.855627302" lastFinishedPulling="2025-12-08 15:02:49.807924901 +0000 UTC m=+990.907931016" observedRunningTime="2025-12-08 15:02:56.867198726 +0000 UTC m=+997.967204831" watchObservedRunningTime="2025-12-08 15:02:56.874229827 +0000 UTC m=+997.974235942" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.891441 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" podStartSLOduration=25.042560101 podStartE2EDuration="31.891420988s" podCreationTimestamp="2025-12-08 15:02:25 +0000 UTC" firstStartedPulling="2025-12-08 15:02:43.046646619 +0000 UTC m=+984.146652734" lastFinishedPulling="2025-12-08 15:02:49.895507506 +0000 UTC m=+990.995513621" observedRunningTime="2025-12-08 15:02:56.887796313 +0000 UTC m=+997.987802418" watchObservedRunningTime="2025-12-08 15:02:56.891420988 +0000 UTC m=+997.991427103" Dec 08 15:02:56 crc kubenswrapper[4894]: I1208 15:02:56.914533 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.728671537 podStartE2EDuration="25.914509964s" podCreationTimestamp="2025-12-08 15:02:31 +0000 UTC" firstStartedPulling="2025-12-08 15:02:43.029215181 +0000 UTC m=+984.129221296" lastFinishedPulling="2025-12-08 15:02:56.215053608 +0000 UTC m=+997.315059723" observedRunningTime="2025-12-08 15:02:56.909664962 +0000 UTC m=+998.009671087" watchObservedRunningTime="2025-12-08 15:02:56.914509964 +0000 UTC m=+998.014516079" Dec 08 15:02:59 crc kubenswrapper[4894]: I1208 15:02:59.946754 4894 generic.go:334] "Generic (PLEG): container finished" podID="1e121dc9-bd10-476c-b28b-06c6dcb09165" containerID="6c23e988b134ff3e80b86cc69f41fa9833cd45e15bf6806c30a6e96245fbd1cf" exitCode=0 Dec 08 15:02:59 crc kubenswrapper[4894]: I1208 15:02:59.946863 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1e121dc9-bd10-476c-b28b-06c6dcb09165","Type":"ContainerDied","Data":"6c23e988b134ff3e80b86cc69f41fa9833cd45e15bf6806c30a6e96245fbd1cf"} Dec 08 15:02:59 crc kubenswrapper[4894]: I1208 15:02:59.949135 4894 generic.go:334] "Generic (PLEG): container finished" podID="755a6729-c028-4206-abc7-5b4bd3825a3e" containerID="86c7359c65ade4a8c61efcae3cdf8e65cacdc9569d2321e180789ea399d4c312" exitCode=0 Dec 08 15:02:59 crc kubenswrapper[4894]: I1208 15:02:59.949205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"755a6729-c028-4206-abc7-5b4bd3825a3e","Type":"ContainerDied","Data":"86c7359c65ade4a8c61efcae3cdf8e65cacdc9569d2321e180789ea399d4c312"} Dec 08 15:02:59 crc kubenswrapper[4894]: I1208 15:02:59.952712 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"0949aa02-5e23-4bd9-9a93-180a59f701ad","Type":"ContainerStarted","Data":"1c7df935a1f3ddc54dee5e03f96d3045c7595821cf5be30ea0e7086f952eb216"} Dec 08 15:02:59 crc kubenswrapper[4894]: I1208 15:02:59.956349 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"31bac65c-93e6-4816-837d-a0f467bb3966","Type":"ContainerStarted","Data":"7373e616bb3a4ae3a45638ef100c7d92ae1d7db3198e3e329a5397937b3f881b"} Dec 08 15:02:59 crc kubenswrapper[4894]: I1208 15:02:59.993323 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.920460488 podStartE2EDuration="24.993282302s" podCreationTimestamp="2025-12-08 15:02:35 +0000 UTC" firstStartedPulling="2025-12-08 15:02:43.347348365 +0000 UTC m=+984.447354480" lastFinishedPulling="2025-12-08 15:02:59.420170179 +0000 UTC m=+1000.520176294" observedRunningTime="2025-12-08 15:02:59.990763933 +0000 UTC m=+1001.090770048" watchObservedRunningTime="2025-12-08 15:02:59.993282302 +0000 UTC m=+1001.093288417" Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.024485 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=8.001930793 podStartE2EDuration="22.024461332s" podCreationTimestamp="2025-12-08 15:02:38 +0000 UTC" firstStartedPulling="2025-12-08 15:02:45.38426886 +0000 UTC m=+986.484274975" lastFinishedPulling="2025-12-08 15:02:59.406799399 +0000 UTC m=+1000.506805514" observedRunningTime="2025-12-08 15:03:00.013468597 +0000 UTC m=+1001.113474712" watchObservedRunningTime="2025-12-08 15:03:00.024461332 +0000 UTC m=+1001.124467447" Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.307748 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.445650 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.482482 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.966654 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1e121dc9-bd10-476c-b28b-06c6dcb09165","Type":"ContainerStarted","Data":"d90d602c4697cb61206eac38e2dcab4904c1c1c24ce3152cfea7a7817cfd8bbf"} Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.969234 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"755a6729-c028-4206-abc7-5b4bd3825a3e","Type":"ContainerStarted","Data":"79f947ad21de55e161fdffcb4dc83bcd65aeba07beed436f82a8d7f04f3dd15e"} Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.971004 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 08 15:03:00 crc kubenswrapper[4894]: I1208 15:03:00.991798 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.107702297 podStartE2EDuration="32.991780261s" podCreationTimestamp="2025-12-08 15:02:28 +0000 UTC" firstStartedPulling="2025-12-08 15:02:43.01106374 +0000 UTC m=+984.111069855" lastFinishedPulling="2025-12-08 15:02:49.895141704 +0000 UTC m=+990.995147819" observedRunningTime="2025-12-08 15:03:00.985157024 +0000 UTC m=+1002.085163159" watchObservedRunningTime="2025-12-08 15:03:00.991780261 +0000 UTC m=+1002.091786366" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.007215 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=26.728464442 podStartE2EDuration="34.007196647s" podCreationTimestamp="2025-12-08 15:02:27 +0000 UTC" firstStartedPulling="2025-12-08 15:02:42.617952068 +0000 UTC m=+983.717958183" lastFinishedPulling="2025-12-08 15:02:49.896684273 +0000 UTC m=+990.996690388" observedRunningTime="2025-12-08 15:03:01.004367908 +0000 UTC m=+1002.104374023" watchObservedRunningTime="2025-12-08 15:03:01.007196647 +0000 UTC m=+1002.107202762" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.014783 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.267044 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ffr6k"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.267627 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" podUID="d164d26b-538c-4474-8c58-c840fecfdba6" containerName="dnsmasq-dns" containerID="cri-o://3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542" gracePeriod=10 Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.278337 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.303645 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-946m4"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.304886 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.306653 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.311231 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.314630 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-946m4"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.322780 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-j22jr"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.324160 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.326254 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.343296 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j22jr"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.351069 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2150dea7-8071-418c-ba2d-20eeb73f2ff8-combined-ca-bundle\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412123 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412168 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412195 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2150dea7-8071-418c-ba2d-20eeb73f2ff8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412218 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150dea7-8071-418c-ba2d-20eeb73f2ff8-config\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412256 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-config\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412303 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rscp8\" (UniqueName: \"kubernetes.io/projected/2150dea7-8071-418c-ba2d-20eeb73f2ff8-kube-api-access-rscp8\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412364 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2150dea7-8071-418c-ba2d-20eeb73f2ff8-ovn-rundir\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412387 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2s9l\" (UniqueName: \"kubernetes.io/projected/d272599f-7a07-480b-b23b-7a7b59eeb933-kube-api-access-s2s9l\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.412437 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2150dea7-8071-418c-ba2d-20eeb73f2ff8-ovs-rundir\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.514793 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.514913 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2150dea7-8071-418c-ba2d-20eeb73f2ff8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.514962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150dea7-8071-418c-ba2d-20eeb73f2ff8-config\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515000 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-config\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515068 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rscp8\" (UniqueName: \"kubernetes.io/projected/2150dea7-8071-418c-ba2d-20eeb73f2ff8-kube-api-access-rscp8\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515170 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2150dea7-8071-418c-ba2d-20eeb73f2ff8-ovn-rundir\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515239 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2s9l\" (UniqueName: \"kubernetes.io/projected/d272599f-7a07-480b-b23b-7a7b59eeb933-kube-api-access-s2s9l\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2150dea7-8071-418c-ba2d-20eeb73f2ff8-ovs-rundir\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515502 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2150dea7-8071-418c-ba2d-20eeb73f2ff8-combined-ca-bundle\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515567 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.515993 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2150dea7-8071-418c-ba2d-20eeb73f2ff8-config\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.516028 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-config\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.516273 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2150dea7-8071-418c-ba2d-20eeb73f2ff8-ovs-rundir\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.516274 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2150dea7-8071-418c-ba2d-20eeb73f2ff8-ovn-rundir\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.516498 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.517057 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.521935 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2150dea7-8071-418c-ba2d-20eeb73f2ff8-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.522402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2150dea7-8071-418c-ba2d-20eeb73f2ff8-combined-ca-bundle\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.537576 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2s9l\" (UniqueName: \"kubernetes.io/projected/d272599f-7a07-480b-b23b-7a7b59eeb933-kube-api-access-s2s9l\") pod \"dnsmasq-dns-7fd796d7df-946m4\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.538107 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rscp8\" (UniqueName: \"kubernetes.io/projected/2150dea7-8071-418c-ba2d-20eeb73f2ff8-kube-api-access-rscp8\") pod \"ovn-controller-metrics-j22jr\" (UID: \"2150dea7-8071-418c-ba2d-20eeb73f2ff8\") " pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.721897 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-946m4"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.721975 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vbxsz"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.723530 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.724837 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.728869 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.737169 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-j22jr" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.740448 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vbxsz"] Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.841717 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.841798 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-config\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.841847 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.841900 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.841923 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcr6g\" (UniqueName: \"kubernetes.io/projected/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-kube-api-access-rcr6g\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.879666 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.943055 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-config\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.943445 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.944469 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.944578 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.944600 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcr6g\" (UniqueName: \"kubernetes.io/projected/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-kube-api-access-rcr6g\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.944685 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.945243 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.946588 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-config\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.945808 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.967923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcr6g\" (UniqueName: \"kubernetes.io/projected/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-kube-api-access-rcr6g\") pod \"dnsmasq-dns-86db49b7ff-vbxsz\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.986695 4894 generic.go:334] "Generic (PLEG): container finished" podID="d164d26b-538c-4474-8c58-c840fecfdba6" containerID="3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542" exitCode=0 Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.986836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" event={"ID":"d164d26b-538c-4474-8c58-c840fecfdba6","Type":"ContainerDied","Data":"3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542"} Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.986883 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" event={"ID":"d164d26b-538c-4474-8c58-c840fecfdba6","Type":"ContainerDied","Data":"6e4c16cdfc113be793315270553370a7e023973a3c0a16ae542ed065d140d1d2"} Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.986916 4894 scope.go:117] "RemoveContainer" containerID="3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542" Dec 08 15:03:01 crc kubenswrapper[4894]: I1208 15:03:01.987047 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ffr6k" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.046302 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-dns-svc\") pod \"d164d26b-538c-4474-8c58-c840fecfdba6\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.046413 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-config\") pod \"d164d26b-538c-4474-8c58-c840fecfdba6\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.046478 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7kxn\" (UniqueName: \"kubernetes.io/projected/d164d26b-538c-4474-8c58-c840fecfdba6-kube-api-access-w7kxn\") pod \"d164d26b-538c-4474-8c58-c840fecfdba6\" (UID: \"d164d26b-538c-4474-8c58-c840fecfdba6\") " Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.056135 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d164d26b-538c-4474-8c58-c840fecfdba6-kube-api-access-w7kxn" (OuterVolumeSpecName: "kube-api-access-w7kxn") pod "d164d26b-538c-4474-8c58-c840fecfdba6" (UID: "d164d26b-538c-4474-8c58-c840fecfdba6"). InnerVolumeSpecName "kube-api-access-w7kxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.061417 4894 scope.go:117] "RemoveContainer" containerID="6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.111043 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.120120 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.120168 4894 scope.go:117] "RemoveContainer" containerID="3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542" Dec 08 15:03:02 crc kubenswrapper[4894]: E1208 15:03:02.120604 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542\": container with ID starting with 3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542 not found: ID does not exist" containerID="3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.120641 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542"} err="failed to get container status \"3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542\": rpc error: code = NotFound desc = could not find container \"3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542\": container with ID starting with 3c497ff2ff8b49b7debf62760fe23f2d1fe5ab0fbb4eb603a1278b9b92d2a542 not found: ID does not exist" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.120659 4894 scope.go:117] "RemoveContainer" containerID="6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a" Dec 08 15:03:02 crc kubenswrapper[4894]: E1208 15:03:02.122984 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a\": container with ID starting with 6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a not found: ID does not exist" containerID="6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.123013 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a"} err="failed to get container status \"6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a\": rpc error: code = NotFound desc = could not find container \"6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a\": container with ID starting with 6b8684da0021807868790e285162a21b02e8a3f988ea07f1e1813ecc601c901a not found: ID does not exist" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.128207 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d164d26b-538c-4474-8c58-c840fecfdba6" (UID: "d164d26b-538c-4474-8c58-c840fecfdba6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.149955 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7kxn\" (UniqueName: \"kubernetes.io/projected/d164d26b-538c-4474-8c58-c840fecfdba6-kube-api-access-w7kxn\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.149989 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.176252 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.191285 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-config" (OuterVolumeSpecName: "config") pod "d164d26b-538c-4474-8c58-c840fecfdba6" (UID: "d164d26b-538c-4474-8c58-c840fecfdba6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.211883 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vbxsz"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.239172 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-j22jr"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.251047 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d164d26b-538c-4474-8c58-c840fecfdba6-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.266330 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-66xxz"] Dec 08 15:03:02 crc kubenswrapper[4894]: E1208 15:03:02.266655 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d164d26b-538c-4474-8c58-c840fecfdba6" containerName="dnsmasq-dns" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.266673 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d164d26b-538c-4474-8c58-c840fecfdba6" containerName="dnsmasq-dns" Dec 08 15:03:02 crc kubenswrapper[4894]: E1208 15:03:02.266685 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d164d26b-538c-4474-8c58-c840fecfdba6" containerName="init" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.266692 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d164d26b-538c-4474-8c58-c840fecfdba6" containerName="init" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.266862 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d164d26b-538c-4474-8c58-c840fecfdba6" containerName="dnsmasq-dns" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.267724 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.274764 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-66xxz"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.314737 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-946m4"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.352791 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.355833 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-config\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.356030 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.356255 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkqnq\" (UniqueName: \"kubernetes.io/projected/eb76bbce-e89a-4955-8d55-8f861419d9af-kube-api-access-vkqnq\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.356585 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-dns-svc\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.365237 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ffr6k"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.371465 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ffr6k"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.457881 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkqnq\" (UniqueName: \"kubernetes.io/projected/eb76bbce-e89a-4955-8d55-8f861419d9af-kube-api-access-vkqnq\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.457925 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-dns-svc\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.457969 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.458028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-config\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.458052 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.458858 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.459350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.459423 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-config\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.460143 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-dns-svc\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.486724 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkqnq\" (UniqueName: \"kubernetes.io/projected/eb76bbce-e89a-4955-8d55-8f861419d9af-kube-api-access-vkqnq\") pod \"dnsmasq-dns-698758b865-66xxz\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.540859 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.542413 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.549199 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.549407 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.549570 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-xm5bd" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.551366 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.552997 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.605647 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.663265 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.663311 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.663386 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhvlg\" (UniqueName: \"kubernetes.io/projected/65f9ac8b-3288-4ea6-abc7-c396fb830f69-kube-api-access-lhvlg\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.663464 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/65f9ac8b-3288-4ea6-abc7-c396fb830f69-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.663516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.663551 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f9ac8b-3288-4ea6-abc7-c396fb830f69-config\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.663577 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65f9ac8b-3288-4ea6-abc7-c396fb830f69-scripts\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.764607 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f9ac8b-3288-4ea6-abc7-c396fb830f69-config\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.764662 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65f9ac8b-3288-4ea6-abc7-c396fb830f69-scripts\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.764760 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.764783 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.764806 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhvlg\" (UniqueName: \"kubernetes.io/projected/65f9ac8b-3288-4ea6-abc7-c396fb830f69-kube-api-access-lhvlg\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.764920 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/65f9ac8b-3288-4ea6-abc7-c396fb830f69-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.764945 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.765702 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f9ac8b-3288-4ea6-abc7-c396fb830f69-config\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.765753 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/65f9ac8b-3288-4ea6-abc7-c396fb830f69-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.766015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65f9ac8b-3288-4ea6-abc7-c396fb830f69-scripts\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.770117 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.770303 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.770637 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vbxsz"] Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.777533 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f9ac8b-3288-4ea6-abc7-c396fb830f69-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: I1208 15:03:02.787102 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhvlg\" (UniqueName: \"kubernetes.io/projected/65f9ac8b-3288-4ea6-abc7-c396fb830f69-kube-api-access-lhvlg\") pod \"ovn-northd-0\" (UID: \"65f9ac8b-3288-4ea6-abc7-c396fb830f69\") " pod="openstack/ovn-northd-0" Dec 08 15:03:02 crc kubenswrapper[4894]: E1208 15:03:02.819563 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd272599f_7a07_480b_b23b_7a7b59eeb933.slice/crio-conmon-c633afc410ff3516333498b58ec2baf65dc1e2e633e4f1ebd9f927c162424f3c.scope\": RecentStats: unable to find data in memory cache]" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.008238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j22jr" event={"ID":"2150dea7-8071-418c-ba2d-20eeb73f2ff8","Type":"ContainerStarted","Data":"7dfa423b912151356120e7503ace9f9707dc2facdadbc21590c723b01a4d4422"} Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.008491 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-j22jr" event={"ID":"2150dea7-8071-418c-ba2d-20eeb73f2ff8","Type":"ContainerStarted","Data":"c23c447136e0b2fd986f8447afbcc9ce159cec3151fac5cbd3fa8bcc4f0af429"} Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.010339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" event={"ID":"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e","Type":"ContainerStarted","Data":"badd4f37bd373aebeeacb53c716178a6113b2b5f586c106c82a6e631681e6f11"} Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.010484 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" podUID="3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" containerName="init" containerID="cri-o://36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553" gracePeriod=10 Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.012482 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.020633 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-946m4" event={"ID":"d272599f-7a07-480b-b23b-7a7b59eeb933","Type":"ContainerDied","Data":"c633afc410ff3516333498b58ec2baf65dc1e2e633e4f1ebd9f927c162424f3c"} Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.020754 4894 generic.go:334] "Generic (PLEG): container finished" podID="d272599f-7a07-480b-b23b-7a7b59eeb933" containerID="c633afc410ff3516333498b58ec2baf65dc1e2e633e4f1ebd9f927c162424f3c" exitCode=0 Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.020917 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-946m4" event={"ID":"d272599f-7a07-480b-b23b-7a7b59eeb933","Type":"ContainerStarted","Data":"2f0306d73e307cf14aa6c4beb3841450fbeceb0680aa48b393ef87edd2d17fb7"} Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.046534 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-j22jr" podStartSLOduration=2.046511268 podStartE2EDuration="2.046511268s" podCreationTimestamp="2025-12-08 15:03:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:03.030218674 +0000 UTC m=+1004.130224779" watchObservedRunningTime="2025-12-08 15:03:03.046511268 +0000 UTC m=+1004.146517383" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.095975 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-66xxz"] Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.219033 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d164d26b-538c-4474-8c58-c840fecfdba6" path="/var/lib/kubelet/pods/d164d26b-538c-4474-8c58-c840fecfdba6/volumes" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.330979 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.342551 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.349865 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.350247 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-ch82b" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.350353 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.351212 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.360865 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.376232 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v4xp\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-kube-api-access-4v4xp\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.376287 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-cache\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.376310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-lock\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.376327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.376357 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.431350 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.460951 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479231 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-sb\") pod \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479275 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-config\") pod \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479306 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-config\") pod \"d272599f-7a07-480b-b23b-7a7b59eeb933\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479347 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-dns-svc\") pod \"d272599f-7a07-480b-b23b-7a7b59eeb933\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479451 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2s9l\" (UniqueName: \"kubernetes.io/projected/d272599f-7a07-480b-b23b-7a7b59eeb933-kube-api-access-s2s9l\") pod \"d272599f-7a07-480b-b23b-7a7b59eeb933\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479520 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcr6g\" (UniqueName: \"kubernetes.io/projected/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-kube-api-access-rcr6g\") pod \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479555 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-dns-svc\") pod \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479614 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-nb\") pod \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\" (UID: \"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479647 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-ovsdbserver-nb\") pod \"d272599f-7a07-480b-b23b-7a7b59eeb933\" (UID: \"d272599f-7a07-480b-b23b-7a7b59eeb933\") " Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.479974 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v4xp\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-kube-api-access-4v4xp\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.480013 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-cache\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.480036 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-lock\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.480053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.480081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.480194 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.480211 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.480246 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift podName:5e8bdec4-ff8a-4db5-b883-68a8b00328ed nodeName:}" failed. No retries permitted until 2025-12-08 15:03:03.980231427 +0000 UTC m=+1005.080237542 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift") pod "swift-storage-0" (UID: "5e8bdec4-ff8a-4db5-b883-68a8b00328ed") : configmap "swift-ring-files" not found Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.480572 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-cache\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.480768 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-lock\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.480838 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.490796 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d272599f-7a07-480b-b23b-7a7b59eeb933-kube-api-access-s2s9l" (OuterVolumeSpecName: "kube-api-access-s2s9l") pod "d272599f-7a07-480b-b23b-7a7b59eeb933" (UID: "d272599f-7a07-480b-b23b-7a7b59eeb933"). InnerVolumeSpecName "kube-api-access-s2s9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.502995 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-kube-api-access-rcr6g" (OuterVolumeSpecName: "kube-api-access-rcr6g") pod "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" (UID: "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e"). InnerVolumeSpecName "kube-api-access-rcr6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.503730 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" (UID: "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.509476 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" (UID: "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.510594 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" (UID: "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.515140 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-config" (OuterVolumeSpecName: "config") pod "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" (UID: "3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.519254 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v4xp\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-kube-api-access-4v4xp\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.520275 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.521350 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d272599f-7a07-480b-b23b-7a7b59eeb933" (UID: "d272599f-7a07-480b-b23b-7a7b59eeb933"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.525441 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d272599f-7a07-480b-b23b-7a7b59eeb933" (UID: "d272599f-7a07-480b-b23b-7a7b59eeb933"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.537551 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-config" (OuterVolumeSpecName: "config") pod "d272599f-7a07-480b-b23b-7a7b59eeb933" (UID: "d272599f-7a07-480b-b23b-7a7b59eeb933"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581292 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2s9l\" (UniqueName: \"kubernetes.io/projected/d272599f-7a07-480b-b23b-7a7b59eeb933-kube-api-access-s2s9l\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581331 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcr6g\" (UniqueName: \"kubernetes.io/projected/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-kube-api-access-rcr6g\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581344 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581357 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581369 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581410 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581421 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581432 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.581496 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d272599f-7a07-480b-b23b-7a7b59eeb933-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.637752 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.843637 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-m4f8h"] Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.844295 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" containerName="init" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.844321 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" containerName="init" Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.844341 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d272599f-7a07-480b-b23b-7a7b59eeb933" containerName="init" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.844350 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d272599f-7a07-480b-b23b-7a7b59eeb933" containerName="init" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.844525 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" containerName="init" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.844547 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d272599f-7a07-480b-b23b-7a7b59eeb933" containerName="init" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.845056 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.853221 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.853332 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.854350 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.876839 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m4f8h"] Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.886691 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-ring-data-devices\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.886742 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-combined-ca-bundle\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.886794 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daf51f62-aa82-474b-b3d4-4746043cf399-etc-swift\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.886973 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-swiftconf\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.887050 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-dispersionconf\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.887119 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z67dz\" (UniqueName: \"kubernetes.io/projected/daf51f62-aa82-474b-b3d4-4746043cf399-kube-api-access-z67dz\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.887202 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-scripts\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989096 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-ring-data-devices\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989169 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-combined-ca-bundle\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989303 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daf51f62-aa82-474b-b3d4-4746043cf399-etc-swift\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989386 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-swiftconf\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989434 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-dispersionconf\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z67dz\" (UniqueName: \"kubernetes.io/projected/daf51f62-aa82-474b-b3d4-4746043cf399-kube-api-access-z67dz\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989555 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-scripts\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.989635 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.989853 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.989875 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 15:03:03 crc kubenswrapper[4894]: E1208 15:03:03.989937 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift podName:5e8bdec4-ff8a-4db5-b883-68a8b00328ed nodeName:}" failed. No retries permitted until 2025-12-08 15:03:04.989910445 +0000 UTC m=+1006.089916560 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift") pod "swift-storage-0" (UID: "5e8bdec4-ff8a-4db5-b883-68a8b00328ed") : configmap "swift-ring-files" not found Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.991402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daf51f62-aa82-474b-b3d4-4746043cf399-etc-swift\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.991615 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-ring-data-devices\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.991633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-scripts\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:03 crc kubenswrapper[4894]: I1208 15:03:03.997489 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-swiftconf\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.003366 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-dispersionconf\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.003454 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-combined-ca-bundle\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.005438 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z67dz\" (UniqueName: \"kubernetes.io/projected/daf51f62-aa82-474b-b3d4-4746043cf399-kube-api-access-z67dz\") pod \"swift-ring-rebalance-m4f8h\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.033124 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"65f9ac8b-3288-4ea6-abc7-c396fb830f69","Type":"ContainerStarted","Data":"0449201c328f85c93972c51c558ac56af131e382fb88da10486a0fd1b2e56932"} Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.036593 4894 generic.go:334] "Generic (PLEG): container finished" podID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerID="10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6" exitCode=0 Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.036693 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-66xxz" event={"ID":"eb76bbce-e89a-4955-8d55-8f861419d9af","Type":"ContainerDied","Data":"10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6"} Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.036858 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-66xxz" event={"ID":"eb76bbce-e89a-4955-8d55-8f861419d9af","Type":"ContainerStarted","Data":"c190ac9a22e355c9873b72fc46b5af193c11c8b79e2030654a3c1ff364caea94"} Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.041039 4894 generic.go:334] "Generic (PLEG): container finished" podID="3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" containerID="36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553" exitCode=0 Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.041397 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.041739 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" event={"ID":"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e","Type":"ContainerDied","Data":"36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553"} Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.041776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vbxsz" event={"ID":"3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e","Type":"ContainerDied","Data":"badd4f37bd373aebeeacb53c716178a6113b2b5f586c106c82a6e631681e6f11"} Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.041795 4894 scope.go:117] "RemoveContainer" containerID="36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.046548 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-946m4" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.047457 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-946m4" event={"ID":"d272599f-7a07-480b-b23b-7a7b59eeb933","Type":"ContainerDied","Data":"2f0306d73e307cf14aa6c4beb3841450fbeceb0680aa48b393ef87edd2d17fb7"} Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.078595 4894 scope.go:117] "RemoveContainer" containerID="36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553" Dec 08 15:03:04 crc kubenswrapper[4894]: E1208 15:03:04.079079 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553\": container with ID starting with 36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553 not found: ID does not exist" containerID="36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.079177 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553"} err="failed to get container status \"36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553\": rpc error: code = NotFound desc = could not find container \"36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553\": container with ID starting with 36726750a762a942d1fea08d6bc4daa5afe4778849fa82e4b0aaccde4c609553 not found: ID does not exist" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.079257 4894 scope.go:117] "RemoveContainer" containerID="c633afc410ff3516333498b58ec2baf65dc1e2e633e4f1ebd9f927c162424f3c" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.123185 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-946m4"] Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.153894 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-946m4"] Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.166761 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vbxsz"] Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.173797 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.174565 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vbxsz"] Dec 08 15:03:04 crc kubenswrapper[4894]: I1208 15:03:04.647337 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m4f8h"] Dec 08 15:03:04 crc kubenswrapper[4894]: W1208 15:03:04.847202 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf51f62_aa82_474b_b3d4_4746043cf399.slice/crio-6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441 WatchSource:0}: Error finding container 6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441: Status 404 returned error can't find the container with id 6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441 Dec 08 15:03:05 crc kubenswrapper[4894]: I1208 15:03:05.007444 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:05 crc kubenswrapper[4894]: E1208 15:03:05.007668 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 15:03:05 crc kubenswrapper[4894]: E1208 15:03:05.007705 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 15:03:05 crc kubenswrapper[4894]: E1208 15:03:05.007762 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift podName:5e8bdec4-ff8a-4db5-b883-68a8b00328ed nodeName:}" failed. No retries permitted until 2025-12-08 15:03:07.007742932 +0000 UTC m=+1008.107749047 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift") pod "swift-storage-0" (UID: "5e8bdec4-ff8a-4db5-b883-68a8b00328ed") : configmap "swift-ring-files" not found Dec 08 15:03:05 crc kubenswrapper[4894]: I1208 15:03:05.053337 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m4f8h" event={"ID":"daf51f62-aa82-474b-b3d4-4746043cf399","Type":"ContainerStarted","Data":"6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441"} Dec 08 15:03:05 crc kubenswrapper[4894]: I1208 15:03:05.056983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-66xxz" event={"ID":"eb76bbce-e89a-4955-8d55-8f861419d9af","Type":"ContainerStarted","Data":"a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632"} Dec 08 15:03:05 crc kubenswrapper[4894]: I1208 15:03:05.057179 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:05 crc kubenswrapper[4894]: I1208 15:03:05.083117 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-66xxz" podStartSLOduration=3.083097172 podStartE2EDuration="3.083097172s" podCreationTimestamp="2025-12-08 15:03:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:05.074909144 +0000 UTC m=+1006.174915259" watchObservedRunningTime="2025-12-08 15:03:05.083097172 +0000 UTC m=+1006.183103287" Dec 08 15:03:05 crc kubenswrapper[4894]: I1208 15:03:05.222926 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e" path="/var/lib/kubelet/pods/3ebc9fa7-ff89-407b-ae6c-823e0aca8e8e/volumes" Dec 08 15:03:05 crc kubenswrapper[4894]: I1208 15:03:05.227753 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d272599f-7a07-480b-b23b-7a7b59eeb933" path="/var/lib/kubelet/pods/d272599f-7a07-480b-b23b-7a7b59eeb933/volumes" Dec 08 15:03:06 crc kubenswrapper[4894]: I1208 15:03:06.120191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"65f9ac8b-3288-4ea6-abc7-c396fb830f69","Type":"ContainerStarted","Data":"aab3d3e22fa617bc0a6a688a080aae4453aba8e59a975d18b868ca26e8b1aa6f"} Dec 08 15:03:06 crc kubenswrapper[4894]: I1208 15:03:06.120563 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"65f9ac8b-3288-4ea6-abc7-c396fb830f69","Type":"ContainerStarted","Data":"cd388dcb43d36310c3dbd6cf047cfe946deb4a7dd86d0d100dfb59f256657049"} Dec 08 15:03:06 crc kubenswrapper[4894]: I1208 15:03:06.154192 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.922185561 podStartE2EDuration="4.154171894s" podCreationTimestamp="2025-12-08 15:03:02 +0000 UTC" firstStartedPulling="2025-12-08 15:03:03.654591239 +0000 UTC m=+1004.754597354" lastFinishedPulling="2025-12-08 15:03:04.886577572 +0000 UTC m=+1005.986583687" observedRunningTime="2025-12-08 15:03:06.147088681 +0000 UTC m=+1007.247094806" watchObservedRunningTime="2025-12-08 15:03:06.154171894 +0000 UTC m=+1007.254178009" Dec 08 15:03:07 crc kubenswrapper[4894]: I1208 15:03:07.042728 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:07 crc kubenswrapper[4894]: E1208 15:03:07.042972 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 15:03:07 crc kubenswrapper[4894]: E1208 15:03:07.043002 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 15:03:07 crc kubenswrapper[4894]: E1208 15:03:07.043072 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift podName:5e8bdec4-ff8a-4db5-b883-68a8b00328ed nodeName:}" failed. No retries permitted until 2025-12-08 15:03:11.043048647 +0000 UTC m=+1012.143054762 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift") pod "swift-storage-0" (UID: "5e8bdec4-ff8a-4db5-b883-68a8b00328ed") : configmap "swift-ring-files" not found Dec 08 15:03:07 crc kubenswrapper[4894]: I1208 15:03:07.127284 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 08 15:03:08 crc kubenswrapper[4894]: I1208 15:03:08.724224 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 08 15:03:08 crc kubenswrapper[4894]: I1208 15:03:08.724550 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 08 15:03:08 crc kubenswrapper[4894]: I1208 15:03:08.807854 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.141077 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m4f8h" event={"ID":"daf51f62-aa82-474b-b3d4-4746043cf399","Type":"ContainerStarted","Data":"20caefc528e727293bfdba568ba21dc7be34c80b65d6cccc014156c89a7a4047"} Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.164439 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-m4f8h" podStartSLOduration=2.714836367 podStartE2EDuration="6.164419027s" podCreationTimestamp="2025-12-08 15:03:03 +0000 UTC" firstStartedPulling="2025-12-08 15:03:04.849907738 +0000 UTC m=+1005.949913853" lastFinishedPulling="2025-12-08 15:03:08.299490398 +0000 UTC m=+1009.399496513" observedRunningTime="2025-12-08 15:03:09.159194713 +0000 UTC m=+1010.259200838" watchObservedRunningTime="2025-12-08 15:03:09.164419027 +0000 UTC m=+1010.264425132" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.214542 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.899302 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b189-account-create-update-cg7wl"] Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.901395 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.907552 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.918743 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b189-account-create-update-cg7wl"] Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.953148 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bgv76"] Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.954255 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.963163 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bgv76"] Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.995035 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x6hf\" (UniqueName: \"kubernetes.io/projected/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-kube-api-access-4x6hf\") pod \"keystone-db-create-bgv76\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.995111 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bktrm\" (UniqueName: \"kubernetes.io/projected/56c30880-9aba-4ace-8bae-51c96dadd06e-kube-api-access-bktrm\") pod \"keystone-b189-account-create-update-cg7wl\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.995437 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-operator-scripts\") pod \"keystone-db-create-bgv76\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:09 crc kubenswrapper[4894]: I1208 15:03:09.995506 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c30880-9aba-4ace-8bae-51c96dadd06e-operator-scripts\") pod \"keystone-b189-account-create-update-cg7wl\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.004337 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.004419 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.090096 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.097852 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x6hf\" (UniqueName: \"kubernetes.io/projected/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-kube-api-access-4x6hf\") pod \"keystone-db-create-bgv76\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.097946 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bktrm\" (UniqueName: \"kubernetes.io/projected/56c30880-9aba-4ace-8bae-51c96dadd06e-kube-api-access-bktrm\") pod \"keystone-b189-account-create-update-cg7wl\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.098071 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-operator-scripts\") pod \"keystone-db-create-bgv76\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.098091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c30880-9aba-4ace-8bae-51c96dadd06e-operator-scripts\") pod \"keystone-b189-account-create-update-cg7wl\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.098756 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c30880-9aba-4ace-8bae-51c96dadd06e-operator-scripts\") pod \"keystone-b189-account-create-update-cg7wl\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.099375 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-operator-scripts\") pod \"keystone-db-create-bgv76\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.120098 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x6hf\" (UniqueName: \"kubernetes.io/projected/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-kube-api-access-4x6hf\") pod \"keystone-db-create-bgv76\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.121067 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bktrm\" (UniqueName: \"kubernetes.io/projected/56c30880-9aba-4ace-8bae-51c96dadd06e-kube-api-access-bktrm\") pod \"keystone-b189-account-create-update-cg7wl\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.216591 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.261963 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-87dgj"] Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.263219 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.268434 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.270125 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.294240 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-87dgj"] Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.367516 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-0cde-account-create-update-gbrxw"] Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.370060 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.374439 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.405105 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0cde-account-create-update-gbrxw"] Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.406100 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28adf25f-022e-4386-b11e-1cd9fec44b3e-operator-scripts\") pod \"placement-db-create-87dgj\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.406183 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwkmr\" (UniqueName: \"kubernetes.io/projected/28adf25f-022e-4386-b11e-1cd9fec44b3e-kube-api-access-qwkmr\") pod \"placement-db-create-87dgj\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.508013 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-operator-scripts\") pod \"placement-0cde-account-create-update-gbrxw\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.508070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28adf25f-022e-4386-b11e-1cd9fec44b3e-operator-scripts\") pod \"placement-db-create-87dgj\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.508134 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncjhc\" (UniqueName: \"kubernetes.io/projected/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-kube-api-access-ncjhc\") pod \"placement-0cde-account-create-update-gbrxw\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.508156 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwkmr\" (UniqueName: \"kubernetes.io/projected/28adf25f-022e-4386-b11e-1cd9fec44b3e-kube-api-access-qwkmr\") pod \"placement-db-create-87dgj\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.509710 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28adf25f-022e-4386-b11e-1cd9fec44b3e-operator-scripts\") pod \"placement-db-create-87dgj\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.526063 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwkmr\" (UniqueName: \"kubernetes.io/projected/28adf25f-022e-4386-b11e-1cd9fec44b3e-kube-api-access-qwkmr\") pod \"placement-db-create-87dgj\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.610230 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-operator-scripts\") pod \"placement-0cde-account-create-update-gbrxw\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.610337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncjhc\" (UniqueName: \"kubernetes.io/projected/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-kube-api-access-ncjhc\") pod \"placement-0cde-account-create-update-gbrxw\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.611264 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-operator-scripts\") pod \"placement-0cde-account-create-update-gbrxw\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.629504 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncjhc\" (UniqueName: \"kubernetes.io/projected/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-kube-api-access-ncjhc\") pod \"placement-0cde-account-create-update-gbrxw\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.723264 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87dgj" Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.735422 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:10 crc kubenswrapper[4894]: W1208 15:03:10.777084 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56c30880_9aba_4ace_8bae_51c96dadd06e.slice/crio-c8a254b184697a7ba27629c317a824a310f95dc852fd982c060e5ae672b6cd1b WatchSource:0}: Error finding container c8a254b184697a7ba27629c317a824a310f95dc852fd982c060e5ae672b6cd1b: Status 404 returned error can't find the container with id c8a254b184697a7ba27629c317a824a310f95dc852fd982c060e5ae672b6cd1b Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.780270 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b189-account-create-update-cg7wl"] Dec 08 15:03:10 crc kubenswrapper[4894]: I1208 15:03:10.933917 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bgv76"] Dec 08 15:03:10 crc kubenswrapper[4894]: W1208 15:03:10.968003 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacc10b3a_43ab_4a08_a981_d6dc24b9b5af.slice/crio-301f65fb4a10b9e440bd271dd3a1e4cf94ddbde4b030e8602aebdcda22645f4c WatchSource:0}: Error finding container 301f65fb4a10b9e440bd271dd3a1e4cf94ddbde4b030e8602aebdcda22645f4c: Status 404 returned error can't find the container with id 301f65fb4a10b9e440bd271dd3a1e4cf94ddbde4b030e8602aebdcda22645f4c Dec 08 15:03:11 crc kubenswrapper[4894]: I1208 15:03:11.119503 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:11 crc kubenswrapper[4894]: E1208 15:03:11.119744 4894 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 08 15:03:11 crc kubenswrapper[4894]: E1208 15:03:11.119757 4894 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 08 15:03:11 crc kubenswrapper[4894]: E1208 15:03:11.119797 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift podName:5e8bdec4-ff8a-4db5-b883-68a8b00328ed nodeName:}" failed. No retries permitted until 2025-12-08 15:03:19.119782728 +0000 UTC m=+1020.219788843 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift") pod "swift-storage-0" (UID: "5e8bdec4-ff8a-4db5-b883-68a8b00328ed") : configmap "swift-ring-files" not found Dec 08 15:03:11 crc kubenswrapper[4894]: I1208 15:03:11.170836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b189-account-create-update-cg7wl" event={"ID":"56c30880-9aba-4ace-8bae-51c96dadd06e","Type":"ContainerStarted","Data":"8ca97fea586b6476c6019ed0179f40675e623cd73290e180b6707f6d74f23592"} Dec 08 15:03:11 crc kubenswrapper[4894]: I1208 15:03:11.171122 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b189-account-create-update-cg7wl" event={"ID":"56c30880-9aba-4ace-8bae-51c96dadd06e","Type":"ContainerStarted","Data":"c8a254b184697a7ba27629c317a824a310f95dc852fd982c060e5ae672b6cd1b"} Dec 08 15:03:11 crc kubenswrapper[4894]: I1208 15:03:11.179021 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bgv76" event={"ID":"acc10b3a-43ab-4a08-a981-d6dc24b9b5af","Type":"ContainerStarted","Data":"301f65fb4a10b9e440bd271dd3a1e4cf94ddbde4b030e8602aebdcda22645f4c"} Dec 08 15:03:11 crc kubenswrapper[4894]: I1208 15:03:11.193060 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b189-account-create-update-cg7wl" podStartSLOduration=2.193040862 podStartE2EDuration="2.193040862s" podCreationTimestamp="2025-12-08 15:03:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:11.189138509 +0000 UTC m=+1012.289144624" watchObservedRunningTime="2025-12-08 15:03:11.193040862 +0000 UTC m=+1012.293046977" Dec 08 15:03:11 crc kubenswrapper[4894]: I1208 15:03:11.282077 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-87dgj"] Dec 08 15:03:11 crc kubenswrapper[4894]: W1208 15:03:11.294891 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28adf25f_022e_4386_b11e_1cd9fec44b3e.slice/crio-7328781067d7ae46edcbfee032c7da0d38e585f65fe962f9ccdb245c11994ae1 WatchSource:0}: Error finding container 7328781067d7ae46edcbfee032c7da0d38e585f65fe962f9ccdb245c11994ae1: Status 404 returned error can't find the container with id 7328781067d7ae46edcbfee032c7da0d38e585f65fe962f9ccdb245c11994ae1 Dec 08 15:03:11 crc kubenswrapper[4894]: I1208 15:03:11.353132 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0cde-account-create-update-gbrxw"] Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.186545 4894 generic.go:334] "Generic (PLEG): container finished" podID="acc10b3a-43ab-4a08-a981-d6dc24b9b5af" containerID="6335e9953b6c8e00d68c28474beaa8b2235a73cfce70533d4408b8c4ac92faf2" exitCode=0 Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.186640 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bgv76" event={"ID":"acc10b3a-43ab-4a08-a981-d6dc24b9b5af","Type":"ContainerDied","Data":"6335e9953b6c8e00d68c28474beaa8b2235a73cfce70533d4408b8c4ac92faf2"} Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.189700 4894 generic.go:334] "Generic (PLEG): container finished" podID="28adf25f-022e-4386-b11e-1cd9fec44b3e" containerID="ac86e1b5e294420010e0f40667ad6ff07cf90eb34032ab45e3720e31a8e6aa79" exitCode=0 Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.189792 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-87dgj" event={"ID":"28adf25f-022e-4386-b11e-1cd9fec44b3e","Type":"ContainerDied","Data":"ac86e1b5e294420010e0f40667ad6ff07cf90eb34032ab45e3720e31a8e6aa79"} Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.189836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-87dgj" event={"ID":"28adf25f-022e-4386-b11e-1cd9fec44b3e","Type":"ContainerStarted","Data":"7328781067d7ae46edcbfee032c7da0d38e585f65fe962f9ccdb245c11994ae1"} Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.191236 4894 generic.go:334] "Generic (PLEG): container finished" podID="56c30880-9aba-4ace-8bae-51c96dadd06e" containerID="8ca97fea586b6476c6019ed0179f40675e623cd73290e180b6707f6d74f23592" exitCode=0 Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.191344 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b189-account-create-update-cg7wl" event={"ID":"56c30880-9aba-4ace-8bae-51c96dadd06e","Type":"ContainerDied","Data":"8ca97fea586b6476c6019ed0179f40675e623cd73290e180b6707f6d74f23592"} Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.192999 4894 generic.go:334] "Generic (PLEG): container finished" podID="ae2689b5-0848-4b5f-8683-ecacf3f46dc8" containerID="683b976da4ad77cc67f2d9d7983262120b45b1f63e3fcea35055d68dfd6653ef" exitCode=0 Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.193052 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0cde-account-create-update-gbrxw" event={"ID":"ae2689b5-0848-4b5f-8683-ecacf3f46dc8","Type":"ContainerDied","Data":"683b976da4ad77cc67f2d9d7983262120b45b1f63e3fcea35055d68dfd6653ef"} Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.193071 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0cde-account-create-update-gbrxw" event={"ID":"ae2689b5-0848-4b5f-8683-ecacf3f46dc8","Type":"ContainerStarted","Data":"d6ef1cf9db202012578130ae22b2aab3a44e17c5920eedfc66f747a20b7e9425"} Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.607950 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.675088 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j85bz"] Dec 08 15:03:12 crc kubenswrapper[4894]: I1208 15:03:12.675330 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" podUID="7619d825-e819-4ab8-810f-584e402aa72c" containerName="dnsmasq-dns" containerID="cri-o://de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593" gracePeriod=10 Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.118474 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.207004 4894 generic.go:334] "Generic (PLEG): container finished" podID="7619d825-e819-4ab8-810f-584e402aa72c" containerID="de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593" exitCode=0 Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.207223 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.218164 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" event={"ID":"7619d825-e819-4ab8-810f-584e402aa72c","Type":"ContainerDied","Data":"de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593"} Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.218213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-j85bz" event={"ID":"7619d825-e819-4ab8-810f-584e402aa72c","Type":"ContainerDied","Data":"82452a7eb2bbb8fbdc941229f93949c7de811f9802b48e6e90bccd8a11a6226a"} Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.218251 4894 scope.go:117] "RemoveContainer" containerID="de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.257953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs9m5\" (UniqueName: \"kubernetes.io/projected/7619d825-e819-4ab8-810f-584e402aa72c-kube-api-access-rs9m5\") pod \"7619d825-e819-4ab8-810f-584e402aa72c\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.258096 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-dns-svc\") pod \"7619d825-e819-4ab8-810f-584e402aa72c\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.258178 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-config\") pod \"7619d825-e819-4ab8-810f-584e402aa72c\" (UID: \"7619d825-e819-4ab8-810f-584e402aa72c\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.270326 4894 scope.go:117] "RemoveContainer" containerID="2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.289247 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7619d825-e819-4ab8-810f-584e402aa72c-kube-api-access-rs9m5" (OuterVolumeSpecName: "kube-api-access-rs9m5") pod "7619d825-e819-4ab8-810f-584e402aa72c" (UID: "7619d825-e819-4ab8-810f-584e402aa72c"). InnerVolumeSpecName "kube-api-access-rs9m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.365063 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs9m5\" (UniqueName: \"kubernetes.io/projected/7619d825-e819-4ab8-810f-584e402aa72c-kube-api-access-rs9m5\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.370197 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7619d825-e819-4ab8-810f-584e402aa72c" (UID: "7619d825-e819-4ab8-810f-584e402aa72c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.381540 4894 scope.go:117] "RemoveContainer" containerID="de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593" Dec 08 15:03:13 crc kubenswrapper[4894]: E1208 15:03:13.382013 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593\": container with ID starting with de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593 not found: ID does not exist" containerID="de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.382047 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593"} err="failed to get container status \"de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593\": rpc error: code = NotFound desc = could not find container \"de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593\": container with ID starting with de95e773e3079b3691feb13a7e72914b0727ba31eb68d7312e2a69616c131593 not found: ID does not exist" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.382067 4894 scope.go:117] "RemoveContainer" containerID="2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5" Dec 08 15:03:13 crc kubenswrapper[4894]: E1208 15:03:13.382274 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5\": container with ID starting with 2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5 not found: ID does not exist" containerID="2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.382299 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5"} err="failed to get container status \"2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5\": rpc error: code = NotFound desc = could not find container \"2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5\": container with ID starting with 2472ed1a94b4432769b572941c22172499acf1604c480dad981fc1369def6dd5 not found: ID does not exist" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.425398 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-config" (OuterVolumeSpecName: "config") pod "7619d825-e819-4ab8-810f-584e402aa72c" (UID: "7619d825-e819-4ab8-810f-584e402aa72c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.467037 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.467312 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7619d825-e819-4ab8-810f-584e402aa72c-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.545896 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j85bz"] Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.552515 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-j85bz"] Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.590860 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.723911 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.734555 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87dgj" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.746392 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.773688 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-operator-scripts\") pod \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.773765 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncjhc\" (UniqueName: \"kubernetes.io/projected/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-kube-api-access-ncjhc\") pod \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\" (UID: \"ae2689b5-0848-4b5f-8683-ecacf3f46dc8\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.774391 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae2689b5-0848-4b5f-8683-ecacf3f46dc8" (UID: "ae2689b5-0848-4b5f-8683-ecacf3f46dc8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.779429 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-kube-api-access-ncjhc" (OuterVolumeSpecName: "kube-api-access-ncjhc") pod "ae2689b5-0848-4b5f-8683-ecacf3f46dc8" (UID: "ae2689b5-0848-4b5f-8683-ecacf3f46dc8"). InnerVolumeSpecName "kube-api-access-ncjhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.875370 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c30880-9aba-4ace-8bae-51c96dadd06e-operator-scripts\") pod \"56c30880-9aba-4ace-8bae-51c96dadd06e\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.875443 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28adf25f-022e-4386-b11e-1cd9fec44b3e-operator-scripts\") pod \"28adf25f-022e-4386-b11e-1cd9fec44b3e\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.875609 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwkmr\" (UniqueName: \"kubernetes.io/projected/28adf25f-022e-4386-b11e-1cd9fec44b3e-kube-api-access-qwkmr\") pod \"28adf25f-022e-4386-b11e-1cd9fec44b3e\" (UID: \"28adf25f-022e-4386-b11e-1cd9fec44b3e\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.875653 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-operator-scripts\") pod \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.875781 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x6hf\" (UniqueName: \"kubernetes.io/projected/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-kube-api-access-4x6hf\") pod \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\" (UID: \"acc10b3a-43ab-4a08-a981-d6dc24b9b5af\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.875809 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bktrm\" (UniqueName: \"kubernetes.io/projected/56c30880-9aba-4ace-8bae-51c96dadd06e-kube-api-access-bktrm\") pod \"56c30880-9aba-4ace-8bae-51c96dadd06e\" (UID: \"56c30880-9aba-4ace-8bae-51c96dadd06e\") " Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.876170 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.876191 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncjhc\" (UniqueName: \"kubernetes.io/projected/ae2689b5-0848-4b5f-8683-ecacf3f46dc8-kube-api-access-ncjhc\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.875794 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56c30880-9aba-4ace-8bae-51c96dadd06e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56c30880-9aba-4ace-8bae-51c96dadd06e" (UID: "56c30880-9aba-4ace-8bae-51c96dadd06e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.876060 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "acc10b3a-43ab-4a08-a981-d6dc24b9b5af" (UID: "acc10b3a-43ab-4a08-a981-d6dc24b9b5af"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.876121 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28adf25f-022e-4386-b11e-1cd9fec44b3e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "28adf25f-022e-4386-b11e-1cd9fec44b3e" (UID: "28adf25f-022e-4386-b11e-1cd9fec44b3e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.878872 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28adf25f-022e-4386-b11e-1cd9fec44b3e-kube-api-access-qwkmr" (OuterVolumeSpecName: "kube-api-access-qwkmr") pod "28adf25f-022e-4386-b11e-1cd9fec44b3e" (UID: "28adf25f-022e-4386-b11e-1cd9fec44b3e"). InnerVolumeSpecName "kube-api-access-qwkmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.879149 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-kube-api-access-4x6hf" (OuterVolumeSpecName: "kube-api-access-4x6hf") pod "acc10b3a-43ab-4a08-a981-d6dc24b9b5af" (UID: "acc10b3a-43ab-4a08-a981-d6dc24b9b5af"). InnerVolumeSpecName "kube-api-access-4x6hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.882963 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56c30880-9aba-4ace-8bae-51c96dadd06e-kube-api-access-bktrm" (OuterVolumeSpecName: "kube-api-access-bktrm") pod "56c30880-9aba-4ace-8bae-51c96dadd06e" (UID: "56c30880-9aba-4ace-8bae-51c96dadd06e"). InnerVolumeSpecName "kube-api-access-bktrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.977478 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x6hf\" (UniqueName: \"kubernetes.io/projected/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-kube-api-access-4x6hf\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.977509 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bktrm\" (UniqueName: \"kubernetes.io/projected/56c30880-9aba-4ace-8bae-51c96dadd06e-kube-api-access-bktrm\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.977521 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c30880-9aba-4ace-8bae-51c96dadd06e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.977529 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/28adf25f-022e-4386-b11e-1cd9fec44b3e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.977538 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwkmr\" (UniqueName: \"kubernetes.io/projected/28adf25f-022e-4386-b11e-1cd9fec44b3e-kube-api-access-qwkmr\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:13 crc kubenswrapper[4894]: I1208 15:03:13.977547 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/acc10b3a-43ab-4a08-a981-d6dc24b9b5af-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.218568 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bgv76" event={"ID":"acc10b3a-43ab-4a08-a981-d6dc24b9b5af","Type":"ContainerDied","Data":"301f65fb4a10b9e440bd271dd3a1e4cf94ddbde4b030e8602aebdcda22645f4c"} Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.218615 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="301f65fb4a10b9e440bd271dd3a1e4cf94ddbde4b030e8602aebdcda22645f4c" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.218582 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bgv76" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.220405 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-87dgj" event={"ID":"28adf25f-022e-4386-b11e-1cd9fec44b3e","Type":"ContainerDied","Data":"7328781067d7ae46edcbfee032c7da0d38e585f65fe962f9ccdb245c11994ae1"} Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.220473 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7328781067d7ae46edcbfee032c7da0d38e585f65fe962f9ccdb245c11994ae1" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.220439 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87dgj" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.222013 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b189-account-create-update-cg7wl" event={"ID":"56c30880-9aba-4ace-8bae-51c96dadd06e","Type":"ContainerDied","Data":"c8a254b184697a7ba27629c317a824a310f95dc852fd982c060e5ae672b6cd1b"} Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.222064 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8a254b184697a7ba27629c317a824a310f95dc852fd982c060e5ae672b6cd1b" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.222138 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b189-account-create-update-cg7wl" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.226521 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0cde-account-create-update-gbrxw" event={"ID":"ae2689b5-0848-4b5f-8683-ecacf3f46dc8","Type":"ContainerDied","Data":"d6ef1cf9db202012578130ae22b2aab3a44e17c5920eedfc66f747a20b7e9425"} Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.226557 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cde-account-create-update-gbrxw" Dec 08 15:03:14 crc kubenswrapper[4894]: I1208 15:03:14.226569 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6ef1cf9db202012578130ae22b2aab3a44e17c5920eedfc66f747a20b7e9425" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.210150 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7619d825-e819-4ab8-810f-584e402aa72c" path="/var/lib/kubelet/pods/7619d825-e819-4ab8-810f-584e402aa72c/volumes" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476020 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-2tlzl"] Dec 08 15:03:15 crc kubenswrapper[4894]: E1208 15:03:15.476633 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2689b5-0848-4b5f-8683-ecacf3f46dc8" containerName="mariadb-account-create-update" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476651 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2689b5-0848-4b5f-8683-ecacf3f46dc8" containerName="mariadb-account-create-update" Dec 08 15:03:15 crc kubenswrapper[4894]: E1208 15:03:15.476671 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7619d825-e819-4ab8-810f-584e402aa72c" containerName="init" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476678 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7619d825-e819-4ab8-810f-584e402aa72c" containerName="init" Dec 08 15:03:15 crc kubenswrapper[4894]: E1208 15:03:15.476695 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28adf25f-022e-4386-b11e-1cd9fec44b3e" containerName="mariadb-database-create" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476703 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="28adf25f-022e-4386-b11e-1cd9fec44b3e" containerName="mariadb-database-create" Dec 08 15:03:15 crc kubenswrapper[4894]: E1208 15:03:15.476716 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c30880-9aba-4ace-8bae-51c96dadd06e" containerName="mariadb-account-create-update" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476725 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c30880-9aba-4ace-8bae-51c96dadd06e" containerName="mariadb-account-create-update" Dec 08 15:03:15 crc kubenswrapper[4894]: E1208 15:03:15.476736 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7619d825-e819-4ab8-810f-584e402aa72c" containerName="dnsmasq-dns" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476743 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7619d825-e819-4ab8-810f-584e402aa72c" containerName="dnsmasq-dns" Dec 08 15:03:15 crc kubenswrapper[4894]: E1208 15:03:15.476756 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acc10b3a-43ab-4a08-a981-d6dc24b9b5af" containerName="mariadb-database-create" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476764 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="acc10b3a-43ab-4a08-a981-d6dc24b9b5af" containerName="mariadb-database-create" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476966 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="28adf25f-022e-4386-b11e-1cd9fec44b3e" containerName="mariadb-database-create" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.476984 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2689b5-0848-4b5f-8683-ecacf3f46dc8" containerName="mariadb-account-create-update" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.478276 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7619d825-e819-4ab8-810f-584e402aa72c" containerName="dnsmasq-dns" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.478305 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="56c30880-9aba-4ace-8bae-51c96dadd06e" containerName="mariadb-account-create-update" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.478336 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="acc10b3a-43ab-4a08-a981-d6dc24b9b5af" containerName="mariadb-database-create" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.479069 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.500500 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr4nh\" (UniqueName: \"kubernetes.io/projected/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-kube-api-access-sr4nh\") pod \"glance-db-create-2tlzl\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.500564 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-operator-scripts\") pod \"glance-db-create-2tlzl\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.501982 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2tlzl"] Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.592880 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-79ea-account-create-update-66lff"] Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.594068 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.596606 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.601730 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr4nh\" (UniqueName: \"kubernetes.io/projected/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-kube-api-access-sr4nh\") pod \"glance-db-create-2tlzl\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.601835 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgkr6\" (UniqueName: \"kubernetes.io/projected/739b42e0-365d-4eff-8724-d0de80c7d29f-kube-api-access-sgkr6\") pod \"glance-79ea-account-create-update-66lff\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.601886 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-operator-scripts\") pod \"glance-db-create-2tlzl\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.601980 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/739b42e0-365d-4eff-8724-d0de80c7d29f-operator-scripts\") pod \"glance-79ea-account-create-update-66lff\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.602846 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-operator-scripts\") pod \"glance-db-create-2tlzl\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.613322 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-79ea-account-create-update-66lff"] Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.640622 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr4nh\" (UniqueName: \"kubernetes.io/projected/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-kube-api-access-sr4nh\") pod \"glance-db-create-2tlzl\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.702965 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgkr6\" (UniqueName: \"kubernetes.io/projected/739b42e0-365d-4eff-8724-d0de80c7d29f-kube-api-access-sgkr6\") pod \"glance-79ea-account-create-update-66lff\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.703064 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/739b42e0-365d-4eff-8724-d0de80c7d29f-operator-scripts\") pod \"glance-79ea-account-create-update-66lff\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.703838 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/739b42e0-365d-4eff-8724-d0de80c7d29f-operator-scripts\") pod \"glance-79ea-account-create-update-66lff\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.727372 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgkr6\" (UniqueName: \"kubernetes.io/projected/739b42e0-365d-4eff-8724-d0de80c7d29f-kube-api-access-sgkr6\") pod \"glance-79ea-account-create-update-66lff\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.798407 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:15 crc kubenswrapper[4894]: I1208 15:03:15.910668 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:16 crc kubenswrapper[4894]: I1208 15:03:16.200387 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-79ea-account-create-update-66lff"] Dec 08 15:03:16 crc kubenswrapper[4894]: W1208 15:03:16.205917 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739b42e0_365d_4eff_8724_d0de80c7d29f.slice/crio-de8dd909b16b3026aca1b250fff4950a617f8b594acfb8f0dd289d41e63d934e WatchSource:0}: Error finding container de8dd909b16b3026aca1b250fff4950a617f8b594acfb8f0dd289d41e63d934e: Status 404 returned error can't find the container with id de8dd909b16b3026aca1b250fff4950a617f8b594acfb8f0dd289d41e63d934e Dec 08 15:03:16 crc kubenswrapper[4894]: I1208 15:03:16.243214 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-79ea-account-create-update-66lff" event={"ID":"739b42e0-365d-4eff-8724-d0de80c7d29f","Type":"ContainerStarted","Data":"de8dd909b16b3026aca1b250fff4950a617f8b594acfb8f0dd289d41e63d934e"} Dec 08 15:03:16 crc kubenswrapper[4894]: I1208 15:03:16.259920 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-2tlzl"] Dec 08 15:03:16 crc kubenswrapper[4894]: W1208 15:03:16.260900 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice/crio-96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9 WatchSource:0}: Error finding container 96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9: Status 404 returned error can't find the container with id 96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9 Dec 08 15:03:17 crc kubenswrapper[4894]: I1208 15:03:17.251667 4894 generic.go:334] "Generic (PLEG): container finished" podID="daf51f62-aa82-474b-b3d4-4746043cf399" containerID="20caefc528e727293bfdba568ba21dc7be34c80b65d6cccc014156c89a7a4047" exitCode=0 Dec 08 15:03:17 crc kubenswrapper[4894]: I1208 15:03:17.251756 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m4f8h" event={"ID":"daf51f62-aa82-474b-b3d4-4746043cf399","Type":"ContainerDied","Data":"20caefc528e727293bfdba568ba21dc7be34c80b65d6cccc014156c89a7a4047"} Dec 08 15:03:17 crc kubenswrapper[4894]: I1208 15:03:17.254256 4894 generic.go:334] "Generic (PLEG): container finished" podID="eb83cfcb-cfb9-4c84-919a-f4408f4962c5" containerID="abd82b1ce872f8717d6a0f43c585d8bfdca536bac15e496575336c6637518624" exitCode=0 Dec 08 15:03:17 crc kubenswrapper[4894]: I1208 15:03:17.254306 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2tlzl" event={"ID":"eb83cfcb-cfb9-4c84-919a-f4408f4962c5","Type":"ContainerDied","Data":"abd82b1ce872f8717d6a0f43c585d8bfdca536bac15e496575336c6637518624"} Dec 08 15:03:17 crc kubenswrapper[4894]: I1208 15:03:17.254340 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2tlzl" event={"ID":"eb83cfcb-cfb9-4c84-919a-f4408f4962c5","Type":"ContainerStarted","Data":"96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9"} Dec 08 15:03:17 crc kubenswrapper[4894]: I1208 15:03:17.255967 4894 generic.go:334] "Generic (PLEG): container finished" podID="739b42e0-365d-4eff-8724-d0de80c7d29f" containerID="02d48033cc2628c0375de65f65e42a41c4853aab4435f51634d3328daffb06ff" exitCode=0 Dec 08 15:03:17 crc kubenswrapper[4894]: I1208 15:03:17.256004 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-79ea-account-create-update-66lff" event={"ID":"739b42e0-365d-4eff-8724-d0de80c7d29f","Type":"ContainerDied","Data":"02d48033cc2628c0375de65f65e42a41c4853aab4435f51634d3328daffb06ff"} Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.065554 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.677392 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.688593 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.697721 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.857087 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-dispersionconf\") pod \"daf51f62-aa82-474b-b3d4-4746043cf399\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.857131 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-scripts\") pod \"daf51f62-aa82-474b-b3d4-4746043cf399\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.857154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr4nh\" (UniqueName: \"kubernetes.io/projected/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-kube-api-access-sr4nh\") pod \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.857969 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z67dz\" (UniqueName: \"kubernetes.io/projected/daf51f62-aa82-474b-b3d4-4746043cf399-kube-api-access-z67dz\") pod \"daf51f62-aa82-474b-b3d4-4746043cf399\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858019 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daf51f62-aa82-474b-b3d4-4746043cf399-etc-swift\") pod \"daf51f62-aa82-474b-b3d4-4746043cf399\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858053 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-combined-ca-bundle\") pod \"daf51f62-aa82-474b-b3d4-4746043cf399\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858177 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgkr6\" (UniqueName: \"kubernetes.io/projected/739b42e0-365d-4eff-8724-d0de80c7d29f-kube-api-access-sgkr6\") pod \"739b42e0-365d-4eff-8724-d0de80c7d29f\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858226 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-operator-scripts\") pod \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\" (UID: \"eb83cfcb-cfb9-4c84-919a-f4408f4962c5\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858262 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/739b42e0-365d-4eff-8724-d0de80c7d29f-operator-scripts\") pod \"739b42e0-365d-4eff-8724-d0de80c7d29f\" (UID: \"739b42e0-365d-4eff-8724-d0de80c7d29f\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858281 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-swiftconf\") pod \"daf51f62-aa82-474b-b3d4-4746043cf399\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858301 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-ring-data-devices\") pod \"daf51f62-aa82-474b-b3d4-4746043cf399\" (UID: \"daf51f62-aa82-474b-b3d4-4746043cf399\") " Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.858950 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eb83cfcb-cfb9-4c84-919a-f4408f4962c5" (UID: "eb83cfcb-cfb9-4c84-919a-f4408f4962c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.859412 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "daf51f62-aa82-474b-b3d4-4746043cf399" (UID: "daf51f62-aa82-474b-b3d4-4746043cf399"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.859771 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/739b42e0-365d-4eff-8724-d0de80c7d29f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "739b42e0-365d-4eff-8724-d0de80c7d29f" (UID: "739b42e0-365d-4eff-8724-d0de80c7d29f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.859902 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daf51f62-aa82-474b-b3d4-4746043cf399-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "daf51f62-aa82-474b-b3d4-4746043cf399" (UID: "daf51f62-aa82-474b-b3d4-4746043cf399"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.863310 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/739b42e0-365d-4eff-8724-d0de80c7d29f-kube-api-access-sgkr6" (OuterVolumeSpecName: "kube-api-access-sgkr6") pod "739b42e0-365d-4eff-8724-d0de80c7d29f" (UID: "739b42e0-365d-4eff-8724-d0de80c7d29f"). InnerVolumeSpecName "kube-api-access-sgkr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.863357 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-kube-api-access-sr4nh" (OuterVolumeSpecName: "kube-api-access-sr4nh") pod "eb83cfcb-cfb9-4c84-919a-f4408f4962c5" (UID: "eb83cfcb-cfb9-4c84-919a-f4408f4962c5"). InnerVolumeSpecName "kube-api-access-sr4nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.865621 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "daf51f62-aa82-474b-b3d4-4746043cf399" (UID: "daf51f62-aa82-474b-b3d4-4746043cf399"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.874159 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf51f62-aa82-474b-b3d4-4746043cf399-kube-api-access-z67dz" (OuterVolumeSpecName: "kube-api-access-z67dz") pod "daf51f62-aa82-474b-b3d4-4746043cf399" (UID: "daf51f62-aa82-474b-b3d4-4746043cf399"). InnerVolumeSpecName "kube-api-access-z67dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.881479 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-scripts" (OuterVolumeSpecName: "scripts") pod "daf51f62-aa82-474b-b3d4-4746043cf399" (UID: "daf51f62-aa82-474b-b3d4-4746043cf399"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.885022 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "daf51f62-aa82-474b-b3d4-4746043cf399" (UID: "daf51f62-aa82-474b-b3d4-4746043cf399"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.894951 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "daf51f62-aa82-474b-b3d4-4746043cf399" (UID: "daf51f62-aa82-474b-b3d4-4746043cf399"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959583 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgkr6\" (UniqueName: \"kubernetes.io/projected/739b42e0-365d-4eff-8724-d0de80c7d29f-kube-api-access-sgkr6\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959619 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959628 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/739b42e0-365d-4eff-8724-d0de80c7d29f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959637 4894 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959646 4894 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959654 4894 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959663 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daf51f62-aa82-474b-b3d4-4746043cf399-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959672 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr4nh\" (UniqueName: \"kubernetes.io/projected/eb83cfcb-cfb9-4c84-919a-f4408f4962c5-kube-api-access-sr4nh\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959682 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z67dz\" (UniqueName: \"kubernetes.io/projected/daf51f62-aa82-474b-b3d4-4746043cf399-kube-api-access-z67dz\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959690 4894 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/daf51f62-aa82-474b-b3d4-4746043cf399-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:18 crc kubenswrapper[4894]: I1208 15:03:18.959698 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf51f62-aa82-474b-b3d4-4746043cf399-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.162533 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.167805 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5e8bdec4-ff8a-4db5-b883-68a8b00328ed-etc-swift\") pod \"swift-storage-0\" (UID: \"5e8bdec4-ff8a-4db5-b883-68a8b00328ed\") " pod="openstack/swift-storage-0" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.271445 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-79ea-account-create-update-66lff" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.271441 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-79ea-account-create-update-66lff" event={"ID":"739b42e0-365d-4eff-8724-d0de80c7d29f","Type":"ContainerDied","Data":"de8dd909b16b3026aca1b250fff4950a617f8b594acfb8f0dd289d41e63d934e"} Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.272057 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de8dd909b16b3026aca1b250fff4950a617f8b594acfb8f0dd289d41e63d934e" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.272715 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m4f8h" event={"ID":"daf51f62-aa82-474b-b3d4-4746043cf399","Type":"ContainerDied","Data":"6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441"} Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.272749 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.272759 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m4f8h" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.273860 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-2tlzl" event={"ID":"eb83cfcb-cfb9-4c84-919a-f4408f4962c5","Type":"ContainerDied","Data":"96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9"} Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.273890 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-2tlzl" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.273891 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.322086 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 08 15:03:19 crc kubenswrapper[4894]: I1208 15:03:19.860694 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.282944 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"86a66edc832db47fe99940defb5b0b6a11f6d9ac01defbf6b38c304e863ee019"} Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.753183 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-htqq9"] Dec 08 15:03:20 crc kubenswrapper[4894]: E1208 15:03:20.753688 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb83cfcb-cfb9-4c84-919a-f4408f4962c5" containerName="mariadb-database-create" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.753701 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb83cfcb-cfb9-4c84-919a-f4408f4962c5" containerName="mariadb-database-create" Dec 08 15:03:20 crc kubenswrapper[4894]: E1208 15:03:20.753720 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf51f62-aa82-474b-b3d4-4746043cf399" containerName="swift-ring-rebalance" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.753725 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf51f62-aa82-474b-b3d4-4746043cf399" containerName="swift-ring-rebalance" Dec 08 15:03:20 crc kubenswrapper[4894]: E1208 15:03:20.753737 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="739b42e0-365d-4eff-8724-d0de80c7d29f" containerName="mariadb-account-create-update" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.753744 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="739b42e0-365d-4eff-8724-d0de80c7d29f" containerName="mariadb-account-create-update" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.753915 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb83cfcb-cfb9-4c84-919a-f4408f4962c5" containerName="mariadb-database-create" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.753927 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="739b42e0-365d-4eff-8724-d0de80c7d29f" containerName="mariadb-account-create-update" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.753940 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf51f62-aa82-474b-b3d4-4746043cf399" containerName="swift-ring-rebalance" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.754433 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.764258 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lg9b9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.764304 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.770456 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-htqq9"] Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.789681 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r6tt\" (UniqueName: \"kubernetes.io/projected/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-kube-api-access-7r6tt\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.789740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-db-sync-config-data\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.789770 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-combined-ca-bundle\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.789907 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-config-data\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.890983 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-config-data\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.891060 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r6tt\" (UniqueName: \"kubernetes.io/projected/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-kube-api-access-7r6tt\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.891089 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-db-sync-config-data\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.891103 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-combined-ca-bundle\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.897220 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-combined-ca-bundle\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.897736 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-db-sync-config-data\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.908011 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r6tt\" (UniqueName: \"kubernetes.io/projected/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-kube-api-access-7r6tt\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:20 crc kubenswrapper[4894]: I1208 15:03:20.909625 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-config-data\") pod \"glance-db-sync-htqq9\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:21 crc kubenswrapper[4894]: I1208 15:03:21.076111 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-htqq9" Dec 08 15:03:21 crc kubenswrapper[4894]: I1208 15:03:21.655112 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-htqq9"] Dec 08 15:03:21 crc kubenswrapper[4894]: W1208 15:03:21.665561 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc5041f7_6cf0_46c8_8fc4_9616005d46b7.slice/crio-e006cf5770d302bc7de70ae01dc3e2c9dbe12a53a58038dcf8a679dfb3544e7d WatchSource:0}: Error finding container e006cf5770d302bc7de70ae01dc3e2c9dbe12a53a58038dcf8a679dfb3544e7d: Status 404 returned error can't find the container with id e006cf5770d302bc7de70ae01dc3e2c9dbe12a53a58038dcf8a679dfb3544e7d Dec 08 15:03:22 crc kubenswrapper[4894]: I1208 15:03:22.314019 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-htqq9" event={"ID":"bc5041f7-6cf0-46c8-8fc4-9616005d46b7","Type":"ContainerStarted","Data":"e006cf5770d302bc7de70ae01dc3e2c9dbe12a53a58038dcf8a679dfb3544e7d"} Dec 08 15:03:22 crc kubenswrapper[4894]: I1208 15:03:22.329093 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"ca4b1ea1426ae11f66593751ae6d9925489d51a53673fa6a4989ee675b848076"} Dec 08 15:03:22 crc kubenswrapper[4894]: I1208 15:03:22.329141 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"8e5b133eadab501bd9d446e7d68f92129c4b1b68ab268b08549acf7db84b9123"} Dec 08 15:03:22 crc kubenswrapper[4894]: I1208 15:03:22.329153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"1d08015b7808ba17443207e00b101c30dbd0b1b3f2cb9a846cb9d64619410e12"} Dec 08 15:03:23 crc kubenswrapper[4894]: E1208 15:03:23.269380 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739b42e0_365d_4eff_8724_d0de80c7d29f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice/crio-96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf51f62_aa82_474b_b3d4_4746043cf399.slice/crio-6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441\": RecentStats: unable to find data in memory cache]" Dec 08 15:03:23 crc kubenswrapper[4894]: I1208 15:03:23.362841 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"dab8d928243e66714ee74b41111b1ae7c84ffe66d29a5c40db7c4b4c6a2f74f1"} Dec 08 15:03:25 crc kubenswrapper[4894]: I1208 15:03:25.380661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"36b940889e4221b12ce247badd99ca3acb70dbdabc0c9b4781c12dada7f600f1"} Dec 08 15:03:25 crc kubenswrapper[4894]: I1208 15:03:25.381231 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"72d9ba905d79e0828434a98f4e1120812f470c42b38f342ef8ee8551cee6242b"} Dec 08 15:03:25 crc kubenswrapper[4894]: I1208 15:03:25.384206 4894 generic.go:334] "Generic (PLEG): container finished" podID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerID="340018ff56d632188be324a9d2eb44cf1163ca716d83f89365cb7acedd19ccc4" exitCode=0 Dec 08 15:03:25 crc kubenswrapper[4894]: I1208 15:03:25.384263 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d911c29-3d60-4990-b6ec-240d1acebc86","Type":"ContainerDied","Data":"340018ff56d632188be324a9d2eb44cf1163ca716d83f89365cb7acedd19ccc4"} Dec 08 15:03:26 crc kubenswrapper[4894]: I1208 15:03:26.414771 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"6f89abae5a45da9ca5d13a1975f0555e7c33116b240ce6ff539a0deda9205733"} Dec 08 15:03:26 crc kubenswrapper[4894]: I1208 15:03:26.415113 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"d70b31319003a2167b02ad55e8cdc8b9046f91548f99d62ebba46b1cafc1c424"} Dec 08 15:03:26 crc kubenswrapper[4894]: I1208 15:03:26.418402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d911c29-3d60-4990-b6ec-240d1acebc86","Type":"ContainerStarted","Data":"d1b87c26301a00fc85781bda83dfabda9df7a6356d16a37d056fb9676ba3b413"} Dec 08 15:03:26 crc kubenswrapper[4894]: I1208 15:03:26.418776 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 08 15:03:26 crc kubenswrapper[4894]: I1208 15:03:26.419840 4894 generic.go:334] "Generic (PLEG): container finished" podID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerID="8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d" exitCode=0 Dec 08 15:03:26 crc kubenswrapper[4894]: I1208 15:03:26.419871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a4dd5876-f3d0-4256-9524-ff3bf39ea28a","Type":"ContainerDied","Data":"8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d"} Dec 08 15:03:26 crc kubenswrapper[4894]: I1208 15:03:26.452433 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=53.782239752 podStartE2EDuration="1m1.452414676s" podCreationTimestamp="2025-12-08 15:02:25 +0000 UTC" firstStartedPulling="2025-12-08 15:02:42.137620403 +0000 UTC m=+983.237626518" lastFinishedPulling="2025-12-08 15:02:49.807795327 +0000 UTC m=+990.907801442" observedRunningTime="2025-12-08 15:03:26.45063 +0000 UTC m=+1027.550636125" watchObservedRunningTime="2025-12-08 15:03:26.452414676 +0000 UTC m=+1027.552420781" Dec 08 15:03:27 crc kubenswrapper[4894]: I1208 15:03:27.432455 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a4dd5876-f3d0-4256-9524-ff3bf39ea28a","Type":"ContainerStarted","Data":"c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1"} Dec 08 15:03:27 crc kubenswrapper[4894]: I1208 15:03:27.433027 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:03:27 crc kubenswrapper[4894]: I1208 15:03:27.560032 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=46.549389908 podStartE2EDuration="1m2.560016636s" podCreationTimestamp="2025-12-08 15:02:25 +0000 UTC" firstStartedPulling="2025-12-08 15:02:33.811593633 +0000 UTC m=+974.911599748" lastFinishedPulling="2025-12-08 15:02:49.822220361 +0000 UTC m=+990.922226476" observedRunningTime="2025-12-08 15:03:27.557196647 +0000 UTC m=+1028.657202772" watchObservedRunningTime="2025-12-08 15:03:27.560016636 +0000 UTC m=+1028.660022751" Dec 08 15:03:29 crc kubenswrapper[4894]: I1208 15:03:29.476088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"df572b910ec9a9e6102777ec0ff309a4cc8496edfbef5f7b80c66ed6cbbbf375"} Dec 08 15:03:30 crc kubenswrapper[4894]: I1208 15:03:30.489351 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"b19cede6736d4a3ae6cf62a51e0db2e39e038ffef7863abfd38aa9ab6019ac68"} Dec 08 15:03:30 crc kubenswrapper[4894]: I1208 15:03:30.659541 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-89p8p" podUID="9d8ef054-d4aa-482b-abad-c961a2a5f2ad" containerName="ovn-controller" probeResult="failure" output=< Dec 08 15:03:30 crc kubenswrapper[4894]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 08 15:03:30 crc kubenswrapper[4894]: > Dec 08 15:03:30 crc kubenswrapper[4894]: I1208 15:03:30.738493 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:03:30 crc kubenswrapper[4894]: I1208 15:03:30.740279 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zxxl4" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.019794 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-89p8p-config-rwlkv"] Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.023604 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.029563 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.049189 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-89p8p-config-rwlkv"] Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.137406 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-log-ovn\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.137479 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-additional-scripts\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.137523 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run-ovn\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.137568 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-scripts\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.137586 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.137618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptkf9\" (UniqueName: \"kubernetes.io/projected/1af562f2-d948-4e87-9edf-4a480f731193-kube-api-access-ptkf9\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243454 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-log-ovn\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243525 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-additional-scripts\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243568 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run-ovn\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243614 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-scripts\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243637 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243675 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptkf9\" (UniqueName: \"kubernetes.io/projected/1af562f2-d948-4e87-9edf-4a480f731193-kube-api-access-ptkf9\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243918 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243945 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run-ovn\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.243956 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-log-ovn\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.244391 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-additional-scripts\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.245932 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-scripts\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.275429 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptkf9\" (UniqueName: \"kubernetes.io/projected/1af562f2-d948-4e87-9edf-4a480f731193-kube-api-access-ptkf9\") pod \"ovn-controller-89p8p-config-rwlkv\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.382114 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.544905 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"a49b91bd0533aadd468b1051f9efc9788107fd2973ed1842363d73a7b4f787ca"} Dec 08 15:03:31 crc kubenswrapper[4894]: I1208 15:03:31.544950 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"7bcc57a6197b9cd2120dfa742cc5c447db7f68008705d84ad2a605216b3ee613"} Dec 08 15:03:32 crc kubenswrapper[4894]: I1208 15:03:32.455132 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-89p8p-config-rwlkv"] Dec 08 15:03:32 crc kubenswrapper[4894]: I1208 15:03:32.558060 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-89p8p-config-rwlkv" event={"ID":"1af562f2-d948-4e87-9edf-4a480f731193","Type":"ContainerStarted","Data":"d0c94cf886076ba0c2d6beaae31d1f8aa5f2620e500f2b89986d17536b819c3c"} Dec 08 15:03:32 crc kubenswrapper[4894]: I1208 15:03:32.566054 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"d4f7496e499f17429803cdebb445adc3e62d48cfa131ef8b07454306957e6e4e"} Dec 08 15:03:32 crc kubenswrapper[4894]: I1208 15:03:32.566096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"265c85cf8498ef39e30001576f4182f3a6fdcbc14e8a1309b782df09fa6e0e05"} Dec 08 15:03:33 crc kubenswrapper[4894]: E1208 15:03:33.472271 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice/crio-96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf51f62_aa82_474b_b3d4_4746043cf399.slice/crio-6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739b42e0_365d_4eff_8724_d0de80c7d29f.slice\": RecentStats: unable to find data in memory cache]" Dec 08 15:03:33 crc kubenswrapper[4894]: I1208 15:03:33.579249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-89p8p-config-rwlkv" event={"ID":"1af562f2-d948-4e87-9edf-4a480f731193","Type":"ContainerStarted","Data":"19bcc9c3e34208a8586fb01e1a19a1ed4f3c02766ce286e57566ebba290aae9a"} Dec 08 15:03:33 crc kubenswrapper[4894]: I1208 15:03:33.605069 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5e8bdec4-ff8a-4db5-b883-68a8b00328ed","Type":"ContainerStarted","Data":"fe402c75df26845230f05cefe8ab7f1dc95b7cc122c1620e8b8a2e9d9d941288"} Dec 08 15:03:33 crc kubenswrapper[4894]: I1208 15:03:33.630918 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-89p8p-config-rwlkv" podStartSLOduration=3.622184184 podStartE2EDuration="3.622184184s" podCreationTimestamp="2025-12-08 15:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:33.61220926 +0000 UTC m=+1034.712215385" watchObservedRunningTime="2025-12-08 15:03:33.622184184 +0000 UTC m=+1034.722190299" Dec 08 15:03:33 crc kubenswrapper[4894]: I1208 15:03:33.710509 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=22.743292248 podStartE2EDuration="31.71048427s" podCreationTimestamp="2025-12-08 15:03:02 +0000 UTC" firstStartedPulling="2025-12-08 15:03:19.87356531 +0000 UTC m=+1020.973571425" lastFinishedPulling="2025-12-08 15:03:28.840757332 +0000 UTC m=+1029.940763447" observedRunningTime="2025-12-08 15:03:33.649928836 +0000 UTC m=+1034.749934951" watchObservedRunningTime="2025-12-08 15:03:33.71048427 +0000 UTC m=+1034.810490385" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.070889 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-mfp9l"] Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.072597 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.076698 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.226571 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-mfp9l"] Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.233903 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.233980 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.234037 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc6lt\" (UniqueName: \"kubernetes.io/projected/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-kube-api-access-fc6lt\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.234064 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.234156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-config\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.234219 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.336288 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc6lt\" (UniqueName: \"kubernetes.io/projected/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-kube-api-access-fc6lt\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.336349 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.336380 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-config\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.336401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.336448 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.336492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.338172 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.338294 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.340885 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.340920 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.341958 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-config\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.382625 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc6lt\" (UniqueName: \"kubernetes.io/projected/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-kube-api-access-fc6lt\") pod \"dnsmasq-dns-77585f5f8c-mfp9l\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.395478 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.617681 4894 generic.go:334] "Generic (PLEG): container finished" podID="1af562f2-d948-4e87-9edf-4a480f731193" containerID="19bcc9c3e34208a8586fb01e1a19a1ed4f3c02766ce286e57566ebba290aae9a" exitCode=0 Dec 08 15:03:34 crc kubenswrapper[4894]: I1208 15:03:34.617763 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-89p8p-config-rwlkv" event={"ID":"1af562f2-d948-4e87-9edf-4a480f731193","Type":"ContainerDied","Data":"19bcc9c3e34208a8586fb01e1a19a1ed4f3c02766ce286e57566ebba290aae9a"} Dec 08 15:03:36 crc kubenswrapper[4894]: I1208 15:03:36.050425 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-89p8p" Dec 08 15:03:36 crc kubenswrapper[4894]: I1208 15:03:36.729474 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 08 15:03:37 crc kubenswrapper[4894]: I1208 15:03:37.058088 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 08 15:03:43 crc kubenswrapper[4894]: E1208 15:03:43.696003 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice/crio-96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf51f62_aa82_474b_b3d4_4746043cf399.slice/crio-6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739b42e0_365d_4eff_8724_d0de80c7d29f.slice\": RecentStats: unable to find data in memory cache]" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.729134 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.833786 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-89p8p-config-rwlkv" event={"ID":"1af562f2-d948-4e87-9edf-4a480f731193","Type":"ContainerDied","Data":"d0c94cf886076ba0c2d6beaae31d1f8aa5f2620e500f2b89986d17536b819c3c"} Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.833839 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0c94cf886076ba0c2d6beaae31d1f8aa5f2620e500f2b89986d17536b819c3c" Dec 08 15:03:46 crc kubenswrapper[4894]: E1208 15:03:46.848769 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 08 15:03:46 crc kubenswrapper[4894]: E1208 15:03:46.849210 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7r6tt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-htqq9_openstack(bc5041f7-6cf0-46c8-8fc4-9616005d46b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:03:46 crc kubenswrapper[4894]: E1208 15:03:46.850399 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-htqq9" podUID="bc5041f7-6cf0-46c8-8fc4-9616005d46b7" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.912279 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945486 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run-ovn\") pod \"1af562f2-d948-4e87-9edf-4a480f731193\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945550 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-log-ovn\") pod \"1af562f2-d948-4e87-9edf-4a480f731193\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945578 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1af562f2-d948-4e87-9edf-4a480f731193" (UID: "1af562f2-d948-4e87-9edf-4a480f731193"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945617 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-additional-scripts\") pod \"1af562f2-d948-4e87-9edf-4a480f731193\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945661 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-scripts\") pod \"1af562f2-d948-4e87-9edf-4a480f731193\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945682 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1af562f2-d948-4e87-9edf-4a480f731193" (UID: "1af562f2-d948-4e87-9edf-4a480f731193"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run\") pod \"1af562f2-d948-4e87-9edf-4a480f731193\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.945911 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptkf9\" (UniqueName: \"kubernetes.io/projected/1af562f2-d948-4e87-9edf-4a480f731193-kube-api-access-ptkf9\") pod \"1af562f2-d948-4e87-9edf-4a480f731193\" (UID: \"1af562f2-d948-4e87-9edf-4a480f731193\") " Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.946381 4894 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.946400 4894 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.946459 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1af562f2-d948-4e87-9edf-4a480f731193" (UID: "1af562f2-d948-4e87-9edf-4a480f731193"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.946490 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run" (OuterVolumeSpecName: "var-run") pod "1af562f2-d948-4e87-9edf-4a480f731193" (UID: "1af562f2-d948-4e87-9edf-4a480f731193"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.946829 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-scripts" (OuterVolumeSpecName: "scripts") pod "1af562f2-d948-4e87-9edf-4a480f731193" (UID: "1af562f2-d948-4e87-9edf-4a480f731193"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:46 crc kubenswrapper[4894]: I1208 15:03:46.963147 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af562f2-d948-4e87-9edf-4a480f731193-kube-api-access-ptkf9" (OuterVolumeSpecName: "kube-api-access-ptkf9") pod "1af562f2-d948-4e87-9edf-4a480f731193" (UID: "1af562f2-d948-4e87-9edf-4a480f731193"). InnerVolumeSpecName "kube-api-access-ptkf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.047562 4894 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1af562f2-d948-4e87-9edf-4a480f731193-var-run\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.047608 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptkf9\" (UniqueName: \"kubernetes.io/projected/1af562f2-d948-4e87-9edf-4a480f731193-kube-api-access-ptkf9\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.047624 4894 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.047635 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1af562f2-d948-4e87-9edf-4a480f731193-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.062892 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.280284 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-mld62"] Dec 08 15:03:47 crc kubenswrapper[4894]: E1208 15:03:47.281243 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af562f2-d948-4e87-9edf-4a480f731193" containerName="ovn-config" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.281264 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af562f2-d948-4e87-9edf-4a480f731193" containerName="ovn-config" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.281436 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af562f2-d948-4e87-9edf-4a480f731193" containerName="ovn-config" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.282147 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.310392 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mld62"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.341261 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d5f6-account-create-update-chx9n"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.342523 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.348254 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.354783 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd7c801c-c3b9-47cd-83ce-6a2de4083655-operator-scripts\") pod \"cinder-db-create-mld62\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.357376 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmcl9\" (UniqueName: \"kubernetes.io/projected/fd7c801c-c3b9-47cd-83ce-6a2de4083655-kube-api-access-jmcl9\") pod \"cinder-db-create-mld62\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.371356 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-tdflp"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.372721 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.393249 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d5f6-account-create-update-chx9n"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.403799 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tdflp"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.452066 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6e42-account-create-update-pbwqs"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.453484 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.457429 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.459094 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab8e2753-f67e-42fa-9374-1a51513d4e62-operator-scripts\") pod \"cinder-d5f6-account-create-update-chx9n\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.459162 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd7c801c-c3b9-47cd-83ce-6a2de4083655-operator-scripts\") pod \"cinder-db-create-mld62\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.459201 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk9wz\" (UniqueName: \"kubernetes.io/projected/ab8e2753-f67e-42fa-9374-1a51513d4e62-kube-api-access-dk9wz\") pod \"cinder-d5f6-account-create-update-chx9n\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.459232 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsk4h\" (UniqueName: \"kubernetes.io/projected/58032ce3-99e4-4cce-a421-7d427cc03cce-kube-api-access-lsk4h\") pod \"barbican-db-create-tdflp\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.459302 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58032ce3-99e4-4cce-a421-7d427cc03cce-operator-scripts\") pod \"barbican-db-create-tdflp\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.459339 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmcl9\" (UniqueName: \"kubernetes.io/projected/fd7c801c-c3b9-47cd-83ce-6a2de4083655-kube-api-access-jmcl9\") pod \"cinder-db-create-mld62\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.460393 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd7c801c-c3b9-47cd-83ce-6a2de4083655-operator-scripts\") pod \"cinder-db-create-mld62\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.460968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6e42-account-create-update-pbwqs"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.490564 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmcl9\" (UniqueName: \"kubernetes.io/projected/fd7c801c-c3b9-47cd-83ce-6a2de4083655-kube-api-access-jmcl9\") pod \"cinder-db-create-mld62\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.495626 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-mfp9l"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.528799 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-stx47"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.530097 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.538672 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-stx47"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.561185 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10d506c-37b1-4046-99df-d106a244fed2-operator-scripts\") pod \"barbican-6e42-account-create-update-pbwqs\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.561328 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab8e2753-f67e-42fa-9374-1a51513d4e62-operator-scripts\") pod \"cinder-d5f6-account-create-update-chx9n\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.561399 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk9wz\" (UniqueName: \"kubernetes.io/projected/ab8e2753-f67e-42fa-9374-1a51513d4e62-kube-api-access-dk9wz\") pod \"cinder-d5f6-account-create-update-chx9n\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.561427 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsk4h\" (UniqueName: \"kubernetes.io/projected/58032ce3-99e4-4cce-a421-7d427cc03cce-kube-api-access-lsk4h\") pod \"barbican-db-create-tdflp\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.561465 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh92r\" (UniqueName: \"kubernetes.io/projected/f10d506c-37b1-4046-99df-d106a244fed2-kube-api-access-jh92r\") pod \"barbican-6e42-account-create-update-pbwqs\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.561500 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58032ce3-99e4-4cce-a421-7d427cc03cce-operator-scripts\") pod \"barbican-db-create-tdflp\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.562135 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab8e2753-f67e-42fa-9374-1a51513d4e62-operator-scripts\") pod \"cinder-d5f6-account-create-update-chx9n\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.569909 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58032ce3-99e4-4cce-a421-7d427cc03cce-operator-scripts\") pod \"barbican-db-create-tdflp\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.585578 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk9wz\" (UniqueName: \"kubernetes.io/projected/ab8e2753-f67e-42fa-9374-1a51513d4e62-kube-api-access-dk9wz\") pod \"cinder-d5f6-account-create-update-chx9n\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.597980 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsk4h\" (UniqueName: \"kubernetes.io/projected/58032ce3-99e4-4cce-a421-7d427cc03cce-kube-api-access-lsk4h\") pod \"barbican-db-create-tdflp\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.617578 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ztqbr"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.621964 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.626922 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dscs7" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.627149 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.627322 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.627542 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.642306 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c7e9-account-create-update-qrjvs"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.642654 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mld62" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.655208 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ztqbr"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.655378 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.659029 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.662651 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10d506c-37b1-4046-99df-d106a244fed2-operator-scripts\") pod \"barbican-6e42-account-create-update-pbwqs\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.662779 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65b6af4f-39e9-4afe-be5c-eaa60088ecde-operator-scripts\") pod \"neutron-db-create-stx47\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.662899 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh92r\" (UniqueName: \"kubernetes.io/projected/f10d506c-37b1-4046-99df-d106a244fed2-kube-api-access-jh92r\") pod \"barbican-6e42-account-create-update-pbwqs\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.662946 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgvxg\" (UniqueName: \"kubernetes.io/projected/65b6af4f-39e9-4afe-be5c-eaa60088ecde-kube-api-access-hgvxg\") pod \"neutron-db-create-stx47\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.668219 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.668491 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10d506c-37b1-4046-99df-d106a244fed2-operator-scripts\") pod \"barbican-6e42-account-create-update-pbwqs\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.677794 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7e9-account-create-update-qrjvs"] Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.688275 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh92r\" (UniqueName: \"kubernetes.io/projected/f10d506c-37b1-4046-99df-d106a244fed2-kube-api-access-jh92r\") pod \"barbican-6e42-account-create-update-pbwqs\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.695044 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.764513 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-operator-scripts\") pod \"neutron-c7e9-account-create-update-qrjvs\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.764892 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65b6af4f-39e9-4afe-be5c-eaa60088ecde-operator-scripts\") pod \"neutron-db-create-stx47\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.764948 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-config-data\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.764985 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh9lm\" (UniqueName: \"kubernetes.io/projected/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-kube-api-access-lh9lm\") pod \"neutron-c7e9-account-create-update-qrjvs\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.765015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgvxg\" (UniqueName: \"kubernetes.io/projected/65b6af4f-39e9-4afe-be5c-eaa60088ecde-kube-api-access-hgvxg\") pod \"neutron-db-create-stx47\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.765054 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-combined-ca-bundle\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.765076 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs29l\" (UniqueName: \"kubernetes.io/projected/473b20c5-eb88-4b48-b198-f828dca0e7c0-kube-api-access-gs29l\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.766013 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65b6af4f-39e9-4afe-be5c-eaa60088ecde-operator-scripts\") pod \"neutron-db-create-stx47\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.782504 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.786677 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgvxg\" (UniqueName: \"kubernetes.io/projected/65b6af4f-39e9-4afe-be5c-eaa60088ecde-kube-api-access-hgvxg\") pod \"neutron-db-create-stx47\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.844766 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-89p8p-config-rwlkv" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.846280 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" event={"ID":"b5f234d1-2387-40b5-bfc9-0b899c6e7b07","Type":"ContainerStarted","Data":"e68896aff1f151fb7ab4bf41adff909ce5419d7da4b625b06f94fcd93933b9aa"} Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.846313 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" event={"ID":"b5f234d1-2387-40b5-bfc9-0b899c6e7b07","Type":"ContainerStarted","Data":"79a7bee0cf82c48542552adf4a1bd327b4a6493396fbaf9a020323b1300b7e0c"} Dec 08 15:03:47 crc kubenswrapper[4894]: E1208 15:03:47.847723 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-htqq9" podUID="bc5041f7-6cf0-46c8-8fc4-9616005d46b7" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.866471 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-combined-ca-bundle\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.866519 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs29l\" (UniqueName: \"kubernetes.io/projected/473b20c5-eb88-4b48-b198-f828dca0e7c0-kube-api-access-gs29l\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.866583 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-operator-scripts\") pod \"neutron-c7e9-account-create-update-qrjvs\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.866642 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-config-data\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.866690 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh9lm\" (UniqueName: \"kubernetes.io/projected/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-kube-api-access-lh9lm\") pod \"neutron-c7e9-account-create-update-qrjvs\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.868097 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-operator-scripts\") pod \"neutron-c7e9-account-create-update-qrjvs\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.872575 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-combined-ca-bundle\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.886549 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-stx47" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.894696 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs29l\" (UniqueName: \"kubernetes.io/projected/473b20c5-eb88-4b48-b198-f828dca0e7c0-kube-api-access-gs29l\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.895029 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh9lm\" (UniqueName: \"kubernetes.io/projected/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-kube-api-access-lh9lm\") pod \"neutron-c7e9-account-create-update-qrjvs\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.897454 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-config-data\") pod \"keystone-db-sync-ztqbr\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:47 crc kubenswrapper[4894]: I1208 15:03:47.953807 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:47.997253 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.061464 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-89p8p-config-rwlkv"] Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.218573 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-89p8p-config-rwlkv"] Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.431845 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d5f6-account-create-update-chx9n"] Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.636122 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-mld62"] Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.888308 4894 generic.go:334] "Generic (PLEG): container finished" podID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerID="e68896aff1f151fb7ab4bf41adff909ce5419d7da4b625b06f94fcd93933b9aa" exitCode=0 Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.888639 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" event={"ID":"b5f234d1-2387-40b5-bfc9-0b899c6e7b07","Type":"ContainerDied","Data":"e68896aff1f151fb7ab4bf41adff909ce5419d7da4b625b06f94fcd93933b9aa"} Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.896025 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d5f6-account-create-update-chx9n" event={"ID":"ab8e2753-f67e-42fa-9374-1a51513d4e62","Type":"ContainerStarted","Data":"077ecace2cc9e820c860d2d8db70dda18705dcf42fde3c7e0d4239a26a2741de"} Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.896070 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d5f6-account-create-update-chx9n" event={"ID":"ab8e2753-f67e-42fa-9374-1a51513d4e62","Type":"ContainerStarted","Data":"00d4d19b2272568e226fb9bc4db1c52189f9021eaa13e8f248a644e5f0544f97"} Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.919149 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mld62" event={"ID":"fd7c801c-c3b9-47cd-83ce-6a2de4083655","Type":"ContainerStarted","Data":"ae3facd589274ebc9dcf97d8d96f4d31915c191790931a4e6b7269ca48a29652"} Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.950972 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d5f6-account-create-update-chx9n" podStartSLOduration=1.95094705 podStartE2EDuration="1.95094705s" podCreationTimestamp="2025-12-08 15:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:48.939652154 +0000 UTC m=+1050.039658269" watchObservedRunningTime="2025-12-08 15:03:48.95094705 +0000 UTC m=+1050.050953165" Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.967693 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6e42-account-create-update-pbwqs"] Dec 08 15:03:48 crc kubenswrapper[4894]: I1208 15:03:48.985509 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-mld62" podStartSLOduration=1.985469965 podStartE2EDuration="1.985469965s" podCreationTimestamp="2025-12-08 15:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:48.977531515 +0000 UTC m=+1050.077537640" watchObservedRunningTime="2025-12-08 15:03:48.985469965 +0000 UTC m=+1050.085476080" Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.153559 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7e9-account-create-update-qrjvs"] Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.176305 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-stx47"] Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.184762 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ztqbr"] Dec 08 15:03:49 crc kubenswrapper[4894]: W1208 15:03:49.190037 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65b6af4f_39e9_4afe_be5c_eaa60088ecde.slice/crio-ba6c6c65c7670d02c0f007bbdf70d3a58b5322274fdfda1e86ab555cf20aaff5 WatchSource:0}: Error finding container ba6c6c65c7670d02c0f007bbdf70d3a58b5322274fdfda1e86ab555cf20aaff5: Status 404 returned error can't find the container with id ba6c6c65c7670d02c0f007bbdf70d3a58b5322274fdfda1e86ab555cf20aaff5 Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.235313 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af562f2-d948-4e87-9edf-4a480f731193" path="/var/lib/kubelet/pods/1af562f2-d948-4e87-9edf-4a480f731193/volumes" Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.236011 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tdflp"] Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.939370 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7e9-account-create-update-qrjvs" event={"ID":"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65","Type":"ContainerStarted","Data":"25d99ea25092e78c89ee8bbdc8bf761f07ba80d32f6477c77a453ac7ec0385b9"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.939636 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7e9-account-create-update-qrjvs" event={"ID":"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65","Type":"ContainerStarted","Data":"1fed664df1dc6f4ec362a5f35790491869a2d537baaa2c50c739844390e67e32"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.942052 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tdflp" event={"ID":"58032ce3-99e4-4cce-a421-7d427cc03cce","Type":"ContainerStarted","Data":"ec96aecf3474d70067e2913cacca6f962cf9daf3dced6deb5dc000469be0e303"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.942085 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tdflp" event={"ID":"58032ce3-99e4-4cce-a421-7d427cc03cce","Type":"ContainerStarted","Data":"d72596a6c09f1eaefa4d9eafa5b78c7933d7011c42ccc0c68fc1c300b123e130"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.943788 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6e42-account-create-update-pbwqs" event={"ID":"f10d506c-37b1-4046-99df-d106a244fed2","Type":"ContainerStarted","Data":"e69f3993e82eea147e8983c144e2d29b670963d1602072fb5e21cfd677a72259"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.943843 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6e42-account-create-update-pbwqs" event={"ID":"f10d506c-37b1-4046-99df-d106a244fed2","Type":"ContainerStarted","Data":"8b78d1ef5232486e0e39987ac3bb3a1b8c5055a0336a9c423816c5bf90d4fa9b"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.945535 4894 generic.go:334] "Generic (PLEG): container finished" podID="ab8e2753-f67e-42fa-9374-1a51513d4e62" containerID="077ecace2cc9e820c860d2d8db70dda18705dcf42fde3c7e0d4239a26a2741de" exitCode=0 Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.945595 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d5f6-account-create-update-chx9n" event={"ID":"ab8e2753-f67e-42fa-9374-1a51513d4e62","Type":"ContainerDied","Data":"077ecace2cc9e820c860d2d8db70dda18705dcf42fde3c7e0d4239a26a2741de"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.947396 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ztqbr" event={"ID":"473b20c5-eb88-4b48-b198-f828dca0e7c0","Type":"ContainerStarted","Data":"9acfbbf29b1e88cbe1a3b05594dc4607dde6a840eeb71243f92b778766d23574"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.948752 4894 generic.go:334] "Generic (PLEG): container finished" podID="fd7c801c-c3b9-47cd-83ce-6a2de4083655" containerID="1448200b24854f7d9e77c123734c59d21f22129bd43dca58f3b808e4d1e980fd" exitCode=0 Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.948803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mld62" event={"ID":"fd7c801c-c3b9-47cd-83ce-6a2de4083655","Type":"ContainerDied","Data":"1448200b24854f7d9e77c123734c59d21f22129bd43dca58f3b808e4d1e980fd"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.950125 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-stx47" event={"ID":"65b6af4f-39e9-4afe-be5c-eaa60088ecde","Type":"ContainerStarted","Data":"f12a4b2aa1ee4b7c4f3373e8167c434c3bb616552e964f4b9e0d722b5a57b9a4"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.950182 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-stx47" event={"ID":"65b6af4f-39e9-4afe-be5c-eaa60088ecde","Type":"ContainerStarted","Data":"ba6c6c65c7670d02c0f007bbdf70d3a58b5322274fdfda1e86ab555cf20aaff5"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.957305 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" event={"ID":"b5f234d1-2387-40b5-bfc9-0b899c6e7b07","Type":"ContainerStarted","Data":"dcb2a8fdc65b7cad7f2ea3d18f923d001adad65a1e124f36950210a9d83159e7"} Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.958190 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.971628 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c7e9-account-create-update-qrjvs" podStartSLOduration=2.971611296 podStartE2EDuration="2.971611296s" podCreationTimestamp="2025-12-08 15:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:49.95837584 +0000 UTC m=+1051.058381955" watchObservedRunningTime="2025-12-08 15:03:49.971611296 +0000 UTC m=+1051.071617411" Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.973174 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-stx47" podStartSLOduration=2.973167155 podStartE2EDuration="2.973167155s" podCreationTimestamp="2025-12-08 15:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:49.971706999 +0000 UTC m=+1051.071713124" watchObservedRunningTime="2025-12-08 15:03:49.973167155 +0000 UTC m=+1051.073173270" Dec 08 15:03:49 crc kubenswrapper[4894]: I1208 15:03:49.997550 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" podStartSLOduration=15.997526681 podStartE2EDuration="15.997526681s" podCreationTimestamp="2025-12-08 15:03:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:49.992442442 +0000 UTC m=+1051.092448557" watchObservedRunningTime="2025-12-08 15:03:49.997526681 +0000 UTC m=+1051.097532796" Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.023725 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-tdflp" podStartSLOduration=3.023704664 podStartE2EDuration="3.023704664s" podCreationTimestamp="2025-12-08 15:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:50.01688333 +0000 UTC m=+1051.116889445" watchObservedRunningTime="2025-12-08 15:03:50.023704664 +0000 UTC m=+1051.123710779" Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.051262 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-6e42-account-create-update-pbwqs" podStartSLOduration=3.05123477 podStartE2EDuration="3.05123477s" podCreationTimestamp="2025-12-08 15:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:03:50.03533261 +0000 UTC m=+1051.135338725" watchObservedRunningTime="2025-12-08 15:03:50.05123477 +0000 UTC m=+1051.151240895" Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.969620 4894 generic.go:334] "Generic (PLEG): container finished" podID="f10d506c-37b1-4046-99df-d106a244fed2" containerID="e69f3993e82eea147e8983c144e2d29b670963d1602072fb5e21cfd677a72259" exitCode=0 Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.969701 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6e42-account-create-update-pbwqs" event={"ID":"f10d506c-37b1-4046-99df-d106a244fed2","Type":"ContainerDied","Data":"e69f3993e82eea147e8983c144e2d29b670963d1602072fb5e21cfd677a72259"} Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.971116 4894 generic.go:334] "Generic (PLEG): container finished" podID="65b6af4f-39e9-4afe-be5c-eaa60088ecde" containerID="f12a4b2aa1ee4b7c4f3373e8167c434c3bb616552e964f4b9e0d722b5a57b9a4" exitCode=0 Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.971145 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-stx47" event={"ID":"65b6af4f-39e9-4afe-be5c-eaa60088ecde","Type":"ContainerDied","Data":"f12a4b2aa1ee4b7c4f3373e8167c434c3bb616552e964f4b9e0d722b5a57b9a4"} Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.972452 4894 generic.go:334] "Generic (PLEG): container finished" podID="e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65" containerID="25d99ea25092e78c89ee8bbdc8bf761f07ba80d32f6477c77a453ac7ec0385b9" exitCode=0 Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.972484 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7e9-account-create-update-qrjvs" event={"ID":"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65","Type":"ContainerDied","Data":"25d99ea25092e78c89ee8bbdc8bf761f07ba80d32f6477c77a453ac7ec0385b9"} Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.973851 4894 generic.go:334] "Generic (PLEG): container finished" podID="58032ce3-99e4-4cce-a421-7d427cc03cce" containerID="ec96aecf3474d70067e2913cacca6f962cf9daf3dced6deb5dc000469be0e303" exitCode=0 Dec 08 15:03:50 crc kubenswrapper[4894]: I1208 15:03:50.973899 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tdflp" event={"ID":"58032ce3-99e4-4cce-a421-7d427cc03cce","Type":"ContainerDied","Data":"ec96aecf3474d70067e2913cacca6f962cf9daf3dced6deb5dc000469be0e303"} Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.425080 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mld62" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.431439 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.481165 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk9wz\" (UniqueName: \"kubernetes.io/projected/ab8e2753-f67e-42fa-9374-1a51513d4e62-kube-api-access-dk9wz\") pod \"ab8e2753-f67e-42fa-9374-1a51513d4e62\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.481260 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd7c801c-c3b9-47cd-83ce-6a2de4083655-operator-scripts\") pod \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.481370 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmcl9\" (UniqueName: \"kubernetes.io/projected/fd7c801c-c3b9-47cd-83ce-6a2de4083655-kube-api-access-jmcl9\") pod \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\" (UID: \"fd7c801c-c3b9-47cd-83ce-6a2de4083655\") " Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.481466 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab8e2753-f67e-42fa-9374-1a51513d4e62-operator-scripts\") pod \"ab8e2753-f67e-42fa-9374-1a51513d4e62\" (UID: \"ab8e2753-f67e-42fa-9374-1a51513d4e62\") " Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.483337 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd7c801c-c3b9-47cd-83ce-6a2de4083655-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fd7c801c-c3b9-47cd-83ce-6a2de4083655" (UID: "fd7c801c-c3b9-47cd-83ce-6a2de4083655"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.483612 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab8e2753-f67e-42fa-9374-1a51513d4e62-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab8e2753-f67e-42fa-9374-1a51513d4e62" (UID: "ab8e2753-f67e-42fa-9374-1a51513d4e62"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.515775 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd7c801c-c3b9-47cd-83ce-6a2de4083655-kube-api-access-jmcl9" (OuterVolumeSpecName: "kube-api-access-jmcl9") pod "fd7c801c-c3b9-47cd-83ce-6a2de4083655" (UID: "fd7c801c-c3b9-47cd-83ce-6a2de4083655"). InnerVolumeSpecName "kube-api-access-jmcl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.519642 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab8e2753-f67e-42fa-9374-1a51513d4e62-kube-api-access-dk9wz" (OuterVolumeSpecName: "kube-api-access-dk9wz") pod "ab8e2753-f67e-42fa-9374-1a51513d4e62" (UID: "ab8e2753-f67e-42fa-9374-1a51513d4e62"). InnerVolumeSpecName "kube-api-access-dk9wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.583242 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk9wz\" (UniqueName: \"kubernetes.io/projected/ab8e2753-f67e-42fa-9374-1a51513d4e62-kube-api-access-dk9wz\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.583276 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fd7c801c-c3b9-47cd-83ce-6a2de4083655-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.583286 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmcl9\" (UniqueName: \"kubernetes.io/projected/fd7c801c-c3b9-47cd-83ce-6a2de4083655-kube-api-access-jmcl9\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.583296 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab8e2753-f67e-42fa-9374-1a51513d4e62-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.990415 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d5f6-account-create-update-chx9n" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.990515 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d5f6-account-create-update-chx9n" event={"ID":"ab8e2753-f67e-42fa-9374-1a51513d4e62","Type":"ContainerDied","Data":"00d4d19b2272568e226fb9bc4db1c52189f9021eaa13e8f248a644e5f0544f97"} Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.990556 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00d4d19b2272568e226fb9bc4db1c52189f9021eaa13e8f248a644e5f0544f97" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.997721 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-mld62" Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.998360 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-mld62" event={"ID":"fd7c801c-c3b9-47cd-83ce-6a2de4083655","Type":"ContainerDied","Data":"ae3facd589274ebc9dcf97d8d96f4d31915c191790931a4e6b7269ca48a29652"} Dec 08 15:03:51 crc kubenswrapper[4894]: I1208 15:03:51.998417 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae3facd589274ebc9dcf97d8d96f4d31915c191790931a4e6b7269ca48a29652" Dec 08 15:03:53 crc kubenswrapper[4894]: E1208 15:03:53.903558 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739b42e0_365d_4eff_8724_d0de80c7d29f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice/crio-96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf51f62_aa82_474b_b3d4_4746043cf399.slice/crio-6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441\": RecentStats: unable to find data in memory cache]" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.397602 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.446732 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-66xxz"] Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.447016 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-66xxz" podUID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerName="dnsmasq-dns" containerID="cri-o://a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632" gracePeriod=10 Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.834068 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.840323 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-stx47" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.852221 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.919676 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.927942 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.942102 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgvxg\" (UniqueName: \"kubernetes.io/projected/65b6af4f-39e9-4afe-be5c-eaa60088ecde-kube-api-access-hgvxg\") pod \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.942164 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh9lm\" (UniqueName: \"kubernetes.io/projected/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-kube-api-access-lh9lm\") pod \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.942230 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65b6af4f-39e9-4afe-be5c-eaa60088ecde-operator-scripts\") pod \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\" (UID: \"65b6af4f-39e9-4afe-be5c-eaa60088ecde\") " Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.942286 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-operator-scripts\") pod \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\" (UID: \"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65\") " Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.942974 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65b6af4f-39e9-4afe-be5c-eaa60088ecde-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65b6af4f-39e9-4afe-be5c-eaa60088ecde" (UID: "65b6af4f-39e9-4afe-be5c-eaa60088ecde"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.942982 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65" (UID: "e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.949958 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b6af4f-39e9-4afe-be5c-eaa60088ecde-kube-api-access-hgvxg" (OuterVolumeSpecName: "kube-api-access-hgvxg") pod "65b6af4f-39e9-4afe-be5c-eaa60088ecde" (UID: "65b6af4f-39e9-4afe-be5c-eaa60088ecde"). InnerVolumeSpecName "kube-api-access-hgvxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:54 crc kubenswrapper[4894]: I1208 15:03:54.952069 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-kube-api-access-lh9lm" (OuterVolumeSpecName: "kube-api-access-lh9lm") pod "e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65" (UID: "e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65"). InnerVolumeSpecName "kube-api-access-lh9lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.023724 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-stx47" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.023772 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-stx47" event={"ID":"65b6af4f-39e9-4afe-be5c-eaa60088ecde","Type":"ContainerDied","Data":"ba6c6c65c7670d02c0f007bbdf70d3a58b5322274fdfda1e86ab555cf20aaff5"} Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.023944 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba6c6c65c7670d02c0f007bbdf70d3a58b5322274fdfda1e86ab555cf20aaff5" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.025496 4894 generic.go:334] "Generic (PLEG): container finished" podID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerID="a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632" exitCode=0 Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.025569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-66xxz" event={"ID":"eb76bbce-e89a-4955-8d55-8f861419d9af","Type":"ContainerDied","Data":"a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632"} Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.025588 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-66xxz" event={"ID":"eb76bbce-e89a-4955-8d55-8f861419d9af","Type":"ContainerDied","Data":"c190ac9a22e355c9873b72fc46b5af193c11c8b79e2030654a3c1ff364caea94"} Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.025607 4894 scope.go:117] "RemoveContainer" containerID="a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.025733 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-66xxz" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.030291 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7e9-account-create-update-qrjvs" event={"ID":"e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65","Type":"ContainerDied","Data":"1fed664df1dc6f4ec362a5f35790491869a2d537baaa2c50c739844390e67e32"} Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.030317 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fed664df1dc6f4ec362a5f35790491869a2d537baaa2c50c739844390e67e32" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.030362 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7e9-account-create-update-qrjvs" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.035631 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tdflp" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.036100 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tdflp" event={"ID":"58032ce3-99e4-4cce-a421-7d427cc03cce","Type":"ContainerDied","Data":"d72596a6c09f1eaefa4d9eafa5b78c7933d7011c42ccc0c68fc1c300b123e130"} Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.036134 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d72596a6c09f1eaefa4d9eafa5b78c7933d7011c42ccc0c68fc1c300b123e130" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.038010 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6e42-account-create-update-pbwqs" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.038155 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6e42-account-create-update-pbwqs" event={"ID":"f10d506c-37b1-4046-99df-d106a244fed2","Type":"ContainerDied","Data":"8b78d1ef5232486e0e39987ac3bb3a1b8c5055a0336a9c423816c5bf90d4fa9b"} Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.038196 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b78d1ef5232486e0e39987ac3bb3a1b8c5055a0336a9c423816c5bf90d4fa9b" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.041909 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ztqbr" event={"ID":"473b20c5-eb88-4b48-b198-f828dca0e7c0","Type":"ContainerStarted","Data":"491d9ef680de2ba394919a36913aee194bc87821c18ce36cd7bff21bf8cec0f5"} Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043074 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkqnq\" (UniqueName: \"kubernetes.io/projected/eb76bbce-e89a-4955-8d55-8f861419d9af-kube-api-access-vkqnq\") pod \"eb76bbce-e89a-4955-8d55-8f861419d9af\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043298 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10d506c-37b1-4046-99df-d106a244fed2-operator-scripts\") pod \"f10d506c-37b1-4046-99df-d106a244fed2\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043343 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsk4h\" (UniqueName: \"kubernetes.io/projected/58032ce3-99e4-4cce-a421-7d427cc03cce-kube-api-access-lsk4h\") pod \"58032ce3-99e4-4cce-a421-7d427cc03cce\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043404 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-sb\") pod \"eb76bbce-e89a-4955-8d55-8f861419d9af\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043497 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-dns-svc\") pod \"eb76bbce-e89a-4955-8d55-8f861419d9af\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043554 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58032ce3-99e4-4cce-a421-7d427cc03cce-operator-scripts\") pod \"58032ce3-99e4-4cce-a421-7d427cc03cce\" (UID: \"58032ce3-99e4-4cce-a421-7d427cc03cce\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043580 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-config\") pod \"eb76bbce-e89a-4955-8d55-8f861419d9af\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043721 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh92r\" (UniqueName: \"kubernetes.io/projected/f10d506c-37b1-4046-99df-d106a244fed2-kube-api-access-jh92r\") pod \"f10d506c-37b1-4046-99df-d106a244fed2\" (UID: \"f10d506c-37b1-4046-99df-d106a244fed2\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043794 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-nb\") pod \"eb76bbce-e89a-4955-8d55-8f861419d9af\" (UID: \"eb76bbce-e89a-4955-8d55-8f861419d9af\") " Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.043897 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f10d506c-37b1-4046-99df-d106a244fed2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f10d506c-37b1-4046-99df-d106a244fed2" (UID: "f10d506c-37b1-4046-99df-d106a244fed2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.044578 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65b6af4f-39e9-4afe-be5c-eaa60088ecde-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.044625 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.044641 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f10d506c-37b1-4046-99df-d106a244fed2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.044653 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgvxg\" (UniqueName: \"kubernetes.io/projected/65b6af4f-39e9-4afe-be5c-eaa60088ecde-kube-api-access-hgvxg\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.044668 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh9lm\" (UniqueName: \"kubernetes.io/projected/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65-kube-api-access-lh9lm\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.046245 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb76bbce-e89a-4955-8d55-8f861419d9af-kube-api-access-vkqnq" (OuterVolumeSpecName: "kube-api-access-vkqnq") pod "eb76bbce-e89a-4955-8d55-8f861419d9af" (UID: "eb76bbce-e89a-4955-8d55-8f861419d9af"). InnerVolumeSpecName "kube-api-access-vkqnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.047148 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58032ce3-99e4-4cce-a421-7d427cc03cce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "58032ce3-99e4-4cce-a421-7d427cc03cce" (UID: "58032ce3-99e4-4cce-a421-7d427cc03cce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.047600 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58032ce3-99e4-4cce-a421-7d427cc03cce-kube-api-access-lsk4h" (OuterVolumeSpecName: "kube-api-access-lsk4h") pod "58032ce3-99e4-4cce-a421-7d427cc03cce" (UID: "58032ce3-99e4-4cce-a421-7d427cc03cce"). InnerVolumeSpecName "kube-api-access-lsk4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.048598 4894 scope.go:117] "RemoveContainer" containerID="10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.052605 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10d506c-37b1-4046-99df-d106a244fed2-kube-api-access-jh92r" (OuterVolumeSpecName: "kube-api-access-jh92r") pod "f10d506c-37b1-4046-99df-d106a244fed2" (UID: "f10d506c-37b1-4046-99df-d106a244fed2"). InnerVolumeSpecName "kube-api-access-jh92r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.064628 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ztqbr" podStartSLOduration=2.640874035 podStartE2EDuration="8.064595856s" podCreationTimestamp="2025-12-08 15:03:47 +0000 UTC" firstStartedPulling="2025-12-08 15:03:49.204927136 +0000 UTC m=+1050.304933251" lastFinishedPulling="2025-12-08 15:03:54.628648957 +0000 UTC m=+1055.728655072" observedRunningTime="2025-12-08 15:03:55.057772481 +0000 UTC m=+1056.157778606" watchObservedRunningTime="2025-12-08 15:03:55.064595856 +0000 UTC m=+1056.164601971" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.088343 4894 scope.go:117] "RemoveContainer" containerID="a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632" Dec 08 15:03:55 crc kubenswrapper[4894]: E1208 15:03:55.088739 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632\": container with ID starting with a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632 not found: ID does not exist" containerID="a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.088788 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632"} err="failed to get container status \"a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632\": rpc error: code = NotFound desc = could not find container \"a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632\": container with ID starting with a4f7d1b067de77879900c6b5b44c63ed4b41304ccbe596f10e5cc9c9907b1632 not found: ID does not exist" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.088811 4894 scope.go:117] "RemoveContainer" containerID="10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6" Dec 08 15:03:55 crc kubenswrapper[4894]: E1208 15:03:55.089334 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6\": container with ID starting with 10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6 not found: ID does not exist" containerID="10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.089405 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6"} err="failed to get container status \"10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6\": rpc error: code = NotFound desc = could not find container \"10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6\": container with ID starting with 10c31c236618dc33faa934a2d353a76be6f288243452fc4540fe91a5d64342c6 not found: ID does not exist" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.090956 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-config" (OuterVolumeSpecName: "config") pod "eb76bbce-e89a-4955-8d55-8f861419d9af" (UID: "eb76bbce-e89a-4955-8d55-8f861419d9af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.095619 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb76bbce-e89a-4955-8d55-8f861419d9af" (UID: "eb76bbce-e89a-4955-8d55-8f861419d9af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.096594 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb76bbce-e89a-4955-8d55-8f861419d9af" (UID: "eb76bbce-e89a-4955-8d55-8f861419d9af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.107403 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb76bbce-e89a-4955-8d55-8f861419d9af" (UID: "eb76bbce-e89a-4955-8d55-8f861419d9af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145327 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145363 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkqnq\" (UniqueName: \"kubernetes.io/projected/eb76bbce-e89a-4955-8d55-8f861419d9af-kube-api-access-vkqnq\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145376 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsk4h\" (UniqueName: \"kubernetes.io/projected/58032ce3-99e4-4cce-a421-7d427cc03cce-kube-api-access-lsk4h\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145385 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145394 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145403 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58032ce3-99e4-4cce-a421-7d427cc03cce-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145411 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb76bbce-e89a-4955-8d55-8f861419d9af-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.145421 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh92r\" (UniqueName: \"kubernetes.io/projected/f10d506c-37b1-4046-99df-d106a244fed2-kube-api-access-jh92r\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.358877 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-66xxz"] Dec 08 15:03:55 crc kubenswrapper[4894]: I1208 15:03:55.368682 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-66xxz"] Dec 08 15:03:57 crc kubenswrapper[4894]: I1208 15:03:57.206221 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb76bbce-e89a-4955-8d55-8f861419d9af" path="/var/lib/kubelet/pods/eb76bbce-e89a-4955-8d55-8f861419d9af/volumes" Dec 08 15:03:58 crc kubenswrapper[4894]: I1208 15:03:58.067183 4894 generic.go:334] "Generic (PLEG): container finished" podID="473b20c5-eb88-4b48-b198-f828dca0e7c0" containerID="491d9ef680de2ba394919a36913aee194bc87821c18ce36cd7bff21bf8cec0f5" exitCode=0 Dec 08 15:03:58 crc kubenswrapper[4894]: I1208 15:03:58.067300 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ztqbr" event={"ID":"473b20c5-eb88-4b48-b198-f828dca0e7c0","Type":"ContainerDied","Data":"491d9ef680de2ba394919a36913aee194bc87821c18ce36cd7bff21bf8cec0f5"} Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.417509 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.522414 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs29l\" (UniqueName: \"kubernetes.io/projected/473b20c5-eb88-4b48-b198-f828dca0e7c0-kube-api-access-gs29l\") pod \"473b20c5-eb88-4b48-b198-f828dca0e7c0\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.522607 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-config-data\") pod \"473b20c5-eb88-4b48-b198-f828dca0e7c0\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.522652 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-combined-ca-bundle\") pod \"473b20c5-eb88-4b48-b198-f828dca0e7c0\" (UID: \"473b20c5-eb88-4b48-b198-f828dca0e7c0\") " Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.532326 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/473b20c5-eb88-4b48-b198-f828dca0e7c0-kube-api-access-gs29l" (OuterVolumeSpecName: "kube-api-access-gs29l") pod "473b20c5-eb88-4b48-b198-f828dca0e7c0" (UID: "473b20c5-eb88-4b48-b198-f828dca0e7c0"). InnerVolumeSpecName "kube-api-access-gs29l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.549466 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "473b20c5-eb88-4b48-b198-f828dca0e7c0" (UID: "473b20c5-eb88-4b48-b198-f828dca0e7c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.579133 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-config-data" (OuterVolumeSpecName: "config-data") pod "473b20c5-eb88-4b48-b198-f828dca0e7c0" (UID: "473b20c5-eb88-4b48-b198-f828dca0e7c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.624771 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.624833 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473b20c5-eb88-4b48-b198-f828dca0e7c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:03:59 crc kubenswrapper[4894]: I1208 15:03:59.624845 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs29l\" (UniqueName: \"kubernetes.io/projected/473b20c5-eb88-4b48-b198-f828dca0e7c0-kube-api-access-gs29l\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.082608 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ztqbr" event={"ID":"473b20c5-eb88-4b48-b198-f828dca0e7c0","Type":"ContainerDied","Data":"9acfbbf29b1e88cbe1a3b05594dc4607dde6a840eeb71243f92b778766d23574"} Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.082645 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9acfbbf29b1e88cbe1a3b05594dc4607dde6a840eeb71243f92b778766d23574" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.082658 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ztqbr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.267767 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-v2snv"] Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268395 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="473b20c5-eb88-4b48-b198-f828dca0e7c0" containerName="keystone-db-sync" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268411 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="473b20c5-eb88-4b48-b198-f828dca0e7c0" containerName="keystone-db-sync" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268430 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58032ce3-99e4-4cce-a421-7d427cc03cce" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268439 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="58032ce3-99e4-4cce-a421-7d427cc03cce" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268457 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268463 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268473 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd7c801c-c3b9-47cd-83ce-6a2de4083655" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268478 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd7c801c-c3b9-47cd-83ce-6a2de4083655" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268490 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab8e2753-f67e-42fa-9374-1a51513d4e62" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268497 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab8e2753-f67e-42fa-9374-1a51513d4e62" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268506 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b6af4f-39e9-4afe-be5c-eaa60088ecde" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268512 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b6af4f-39e9-4afe-be5c-eaa60088ecde" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268530 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerName="init" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268536 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerName="init" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268553 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10d506c-37b1-4046-99df-d106a244fed2" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268559 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10d506c-37b1-4046-99df-d106a244fed2" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: E1208 15:04:00.268570 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerName="dnsmasq-dns" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268576 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerName="dnsmasq-dns" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268728 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd7c801c-c3b9-47cd-83ce-6a2de4083655" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268742 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb76bbce-e89a-4955-8d55-8f861419d9af" containerName="dnsmasq-dns" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268750 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="473b20c5-eb88-4b48-b198-f828dca0e7c0" containerName="keystone-db-sync" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268759 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10d506c-37b1-4046-99df-d106a244fed2" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268772 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b6af4f-39e9-4afe-be5c-eaa60088ecde" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268783 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268792 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="58032ce3-99e4-4cce-a421-7d427cc03cce" containerName="mariadb-database-create" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.268800 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab8e2753-f67e-42fa-9374-1a51513d4e62" containerName="mariadb-account-create-update" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.269685 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.288029 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-v2snv"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.388946 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5tvxs"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.390095 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.403352 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.403530 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.403578 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.403731 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.403844 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5tvxs"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.404055 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dscs7" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.440882 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.440969 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqnzf\" (UniqueName: \"kubernetes.io/projected/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-kube-api-access-bqnzf\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.441027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.441063 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-svc\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.441102 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.441198 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-config\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.509738 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7nxjg"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.510943 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.516350 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.516674 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.516809 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zjln5" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.526725 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7nxjg"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.534545 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-749798c7fc-c62cr"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.536255 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.540198 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.540281 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.540372 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.540481 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-lk5zw" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542505 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-credential-keys\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542579 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-scripts\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542623 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-combined-ca-bundle\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542651 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqnzf\" (UniqueName: \"kubernetes.io/projected/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-kube-api-access-bqnzf\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542684 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542710 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-svc\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542754 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4sxm\" (UniqueName: \"kubernetes.io/projected/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-kube-api-access-b4sxm\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542798 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-fernet-keys\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542835 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-config\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.542851 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-config-data\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.543743 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.544725 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.545418 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.546126 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-config\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.547052 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-svc\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.589433 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-749798c7fc-c62cr"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.604728 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqnzf\" (UniqueName: \"kubernetes.io/projected/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-kube-api-access-bqnzf\") pod \"dnsmasq-dns-55fff446b9-v2snv\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.605226 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647006 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-scripts\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-combined-ca-bundle\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647072 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hbvt\" (UniqueName: \"kubernetes.io/projected/06177501-5baa-49e0-94bd-157c6f079c05-kube-api-access-9hbvt\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-db-sync-config-data\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647126 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-config-data\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647157 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/717a577a-7166-44b5-bc64-40553dfe66f8-horizon-secret-key\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647186 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717a577a-7166-44b5-bc64-40553dfe66f8-logs\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647207 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4sxm\" (UniqueName: \"kubernetes.io/projected/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-kube-api-access-b4sxm\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-combined-ca-bundle\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647257 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqnj2\" (UniqueName: \"kubernetes.io/projected/717a577a-7166-44b5-bc64-40553dfe66f8-kube-api-access-qqnj2\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647290 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-fernet-keys\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647307 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-config-data\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06177501-5baa-49e0-94bd-157c6f079c05-etc-machine-id\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647344 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-credential-keys\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647365 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-config-data\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647398 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-scripts\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.647414 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-scripts\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.662504 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-scripts\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.663950 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.665837 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-config-data\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.666433 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.667660 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-fernet-keys\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.668092 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-credential-keys\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.671473 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-combined-ca-bundle\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.680178 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.690262 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.690440 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.713583 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4sxm\" (UniqueName: \"kubernetes.io/projected/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-kube-api-access-b4sxm\") pod \"keystone-bootstrap-5tvxs\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.723217 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.725504 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kzgzh"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.726926 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.749987 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kzgzh"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.755865 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-combined-ca-bundle\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.755909 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqnj2\" (UniqueName: \"kubernetes.io/projected/717a577a-7166-44b5-bc64-40553dfe66f8-kube-api-access-qqnj2\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.755962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06177501-5baa-49e0-94bd-157c6f079c05-etc-machine-id\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756002 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-config-data\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756043 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-scripts\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756072 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-scripts\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756098 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hbvt\" (UniqueName: \"kubernetes.io/projected/06177501-5baa-49e0-94bd-157c6f079c05-kube-api-access-9hbvt\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-db-sync-config-data\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756148 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-config-data\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/717a577a-7166-44b5-bc64-40553dfe66f8-horizon-secret-key\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756205 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717a577a-7166-44b5-bc64-40553dfe66f8-logs\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.756729 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717a577a-7166-44b5-bc64-40553dfe66f8-logs\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.757241 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-scripts\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.764546 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.764704 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.764997 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xr6bk" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.765194 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06177501-5baa-49e0-94bd-157c6f079c05-etc-machine-id\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.766399 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-config-data\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.768451 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-7b8gb"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.770998 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.773658 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/717a577a-7166-44b5-bc64-40553dfe66f8-horizon-secret-key\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.774834 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-scripts\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.781362 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-config-data\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.787600 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.787779 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7b8gb"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.791933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-combined-ca-bundle\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.803405 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-db-sync-config-data\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.820766 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-h78m5" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.825344 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hbvt\" (UniqueName: \"kubernetes.io/projected/06177501-5baa-49e0-94bd-157c6f079c05-kube-api-access-9hbvt\") pod \"cinder-db-sync-7nxjg\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.833937 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqnj2\" (UniqueName: \"kubernetes.io/projected/717a577a-7166-44b5-bc64-40553dfe66f8-kube-api-access-qqnj2\") pod \"horizon-749798c7fc-c62cr\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.841588 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.845215 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c4bf45b45-vmwph"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.846614 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.858804 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-combined-ca-bundle\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.858966 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-log-httpd\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859012 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prfk6\" (UniqueName: \"kubernetes.io/projected/72ff9384-4eab-4600-a3e7-448b15471c6e-kube-api-access-prfk6\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859061 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-config\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859111 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-run-httpd\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859212 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-scripts\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859261 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75k8k\" (UniqueName: \"kubernetes.io/projected/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-kube-api-access-75k8k\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.859292 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-config-data\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.877750 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.893207 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c4bf45b45-vmwph"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.906136 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-v2snv"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.935613 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-g4klr"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.936744 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.942594 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.944039 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.948729 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-x2x98" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.957880 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-2mjnx"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.959595 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968635 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-scripts\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5z2l\" (UniqueName: \"kubernetes.io/projected/85f57102-ceaf-484a-a0b3-c1ecf324339d-kube-api-access-g5z2l\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968699 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-scripts\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968722 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968738 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75k8k\" (UniqueName: \"kubernetes.io/projected/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-kube-api-access-75k8k\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968754 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-combined-ca-bundle\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968772 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-config-data\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968788 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-config-data\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.968807 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-scripts\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982131 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-combined-ca-bundle\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57vq2\" (UniqueName: \"kubernetes.io/projected/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-kube-api-access-57vq2\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982268 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqmmq\" (UniqueName: \"kubernetes.io/projected/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-kube-api-access-nqmmq\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982292 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-log-httpd\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982315 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-config\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982351 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prfk6\" (UniqueName: \"kubernetes.io/projected/72ff9384-4eab-4600-a3e7-448b15471c6e-kube-api-access-prfk6\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982432 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znx6g\" (UniqueName: \"kubernetes.io/projected/195700ca-d25a-470a-a6c4-73d4f4133b95-kube-api-access-znx6g\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982455 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195700ca-d25a-470a-a6c4-73d4f4133b95-logs\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982481 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-config\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982507 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-horizon-secret-key\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982542 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982593 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-run-httpd\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982633 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-logs\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982655 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982677 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982710 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-combined-ca-bundle\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982746 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-config-data\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982763 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.982787 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-db-sync-config-data\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.986036 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-log-httpd\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.986897 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-g4klr"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.987528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-run-httpd\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.990405 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.990532 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-combined-ca-bundle\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.992923 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-2mjnx"] Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.993280 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-config-data\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.994698 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-scripts\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:00 crc kubenswrapper[4894]: I1208 15:04:00.999417 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75k8k\" (UniqueName: \"kubernetes.io/projected/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-kube-api-access-75k8k\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.012426 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-config\") pod \"neutron-db-sync-kzgzh\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.014633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089370 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-scripts\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089657 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5z2l\" (UniqueName: \"kubernetes.io/projected/85f57102-ceaf-484a-a0b3-c1ecf324339d-kube-api-access-g5z2l\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089686 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-combined-ca-bundle\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089706 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-config-data\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089734 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-scripts\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089764 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57vq2\" (UniqueName: \"kubernetes.io/projected/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-kube-api-access-57vq2\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089799 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqmmq\" (UniqueName: \"kubernetes.io/projected/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-kube-api-access-nqmmq\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089829 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-config\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089857 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089890 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znx6g\" (UniqueName: \"kubernetes.io/projected/195700ca-d25a-470a-a6c4-73d4f4133b95-kube-api-access-znx6g\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089907 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195700ca-d25a-470a-a6c4-73d4f4133b95-logs\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089926 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-horizon-secret-key\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089948 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089980 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-logs\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.089996 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.090012 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-combined-ca-bundle\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.090031 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-config-data\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.090048 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.090065 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-db-sync-config-data\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.091186 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.100608 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-scripts\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.101419 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195700ca-d25a-470a-a6c4-73d4f4133b95-logs\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.101651 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-config\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.102105 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-db-sync-config-data\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.102937 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.103406 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.105485 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-combined-ca-bundle\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.105609 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prfk6\" (UniqueName: \"kubernetes.io/projected/72ff9384-4eab-4600-a3e7-448b15471c6e-kube-api-access-prfk6\") pod \"ceilometer-0\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " pod="openstack/ceilometer-0" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.106148 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.113709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-logs\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.114589 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-config-data\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.114889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-scripts\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.123556 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-horizon-secret-key\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.140706 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57vq2\" (UniqueName: \"kubernetes.io/projected/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-kube-api-access-57vq2\") pod \"dnsmasq-dns-76fcf4b695-2mjnx\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.148544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znx6g\" (UniqueName: \"kubernetes.io/projected/195700ca-d25a-470a-a6c4-73d4f4133b95-kube-api-access-znx6g\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.148911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-combined-ca-bundle\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.150683 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-config-data\") pod \"placement-db-sync-g4klr\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.200231 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.202967 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqmmq\" (UniqueName: \"kubernetes.io/projected/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-kube-api-access-nqmmq\") pod \"horizon-c4bf45b45-vmwph\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.203595 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5z2l\" (UniqueName: \"kubernetes.io/projected/85f57102-ceaf-484a-a0b3-c1ecf324339d-kube-api-access-g5z2l\") pod \"barbican-db-sync-7b8gb\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.216917 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.258735 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.305038 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.344693 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.392384 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.636019 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-v2snv"] Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.686175 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5tvxs"] Dec 08 15:04:01 crc kubenswrapper[4894]: I1208 15:04:01.963407 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7nxjg"] Dec 08 15:04:01 crc kubenswrapper[4894]: W1208 15:04:01.973204 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06177501_5baa_49e0_94bd_157c6f079c05.slice/crio-45260764b5a80ebb2fd1b7736a8a527c6488e09823c9208279392d2bda1db5ec WatchSource:0}: Error finding container 45260764b5a80ebb2fd1b7736a8a527c6488e09823c9208279392d2bda1db5ec: Status 404 returned error can't find the container with id 45260764b5a80ebb2fd1b7736a8a527c6488e09823c9208279392d2bda1db5ec Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.138662 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5tvxs" event={"ID":"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29","Type":"ContainerStarted","Data":"548607c0f720255da6e0f5953d7edf31caf09aea918dc6d6c6ddd0e5569d351a"} Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.138935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5tvxs" event={"ID":"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29","Type":"ContainerStarted","Data":"4e0281815a25957438232fc2a93f474087b4e825db050d7d325002f9613e124b"} Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.141211 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7nxjg" event={"ID":"06177501-5baa-49e0-94bd-157c6f079c05","Type":"ContainerStarted","Data":"45260764b5a80ebb2fd1b7736a8a527c6488e09823c9208279392d2bda1db5ec"} Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.144367 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" event={"ID":"6a8d9046-faf2-413c-afc9-3c0eb3076b8e","Type":"ContainerStarted","Data":"ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051"} Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.144414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" event={"ID":"6a8d9046-faf2-413c-afc9-3c0eb3076b8e","Type":"ContainerStarted","Data":"ff2daef663aa9e837f3da0611539e04a23b462e001fee8a217023083e916317f"} Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.144543 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" podUID="6a8d9046-faf2-413c-afc9-3c0eb3076b8e" containerName="init" containerID="cri-o://ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051" gracePeriod=10 Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.160469 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5tvxs" podStartSLOduration=2.16044765 podStartE2EDuration="2.16044765s" podCreationTimestamp="2025-12-08 15:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:02.160332836 +0000 UTC m=+1063.260338971" watchObservedRunningTime="2025-12-08 15:04:02.16044765 +0000 UTC m=+1063.260453765" Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.428118 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.453253 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-2mjnx"] Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.478108 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7b8gb"] Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.493038 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-749798c7fc-c62cr"] Dec 08 15:04:02 crc kubenswrapper[4894]: W1208 15:04:02.497569 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfdaa6a7_4991_4cc0_b19c_1d2e0f067f4f.slice/crio-2130ce948bc6cbbd5163025152e5ebcf953cf54295111470dd8439c64e305824 WatchSource:0}: Error finding container 2130ce948bc6cbbd5163025152e5ebcf953cf54295111470dd8439c64e305824: Status 404 returned error can't find the container with id 2130ce948bc6cbbd5163025152e5ebcf953cf54295111470dd8439c64e305824 Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.500731 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kzgzh"] Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.769490 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-g4klr"] Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.778448 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c4bf45b45-vmwph"] Dec 08 15:04:02 crc kubenswrapper[4894]: W1208 15:04:02.792534 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod586511a0_9e32_4c9b_bbff_1cb74f0a0cc2.slice/crio-2e09635f053889c3bcaad95c49709371e46226fb02382c4572c0e71477fc90b7 WatchSource:0}: Error finding container 2e09635f053889c3bcaad95c49709371e46226fb02382c4572c0e71477fc90b7: Status 404 returned error can't find the container with id 2e09635f053889c3bcaad95c49709371e46226fb02382c4572c0e71477fc90b7 Dec 08 15:04:02 crc kubenswrapper[4894]: W1208 15:04:02.794891 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod195700ca_d25a_470a_a6c4_73d4f4133b95.slice/crio-02412146543e671f6f92e070e4faa0ce4e09c3809a84e4eeef1659d66ef63d52 WatchSource:0}: Error finding container 02412146543e671f6f92e070e4faa0ce4e09c3809a84e4eeef1659d66ef63d52: Status 404 returned error can't find the container with id 02412146543e671f6f92e070e4faa0ce4e09c3809a84e4eeef1659d66ef63d52 Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.899867 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.939195 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-749798c7fc-c62cr"] Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.987658 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79476f47f-8zstk"] Dec 08 15:04:02 crc kubenswrapper[4894]: E1208 15:04:02.988095 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8d9046-faf2-413c-afc9-3c0eb3076b8e" containerName="init" Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.988110 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8d9046-faf2-413c-afc9-3c0eb3076b8e" containerName="init" Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.988305 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8d9046-faf2-413c-afc9-3c0eb3076b8e" containerName="init" Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.989174 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.992222 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-config\") pod \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.992278 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-nb\") pod \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.992345 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-sb\") pod \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.992372 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-swift-storage-0\") pod \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.992435 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqnzf\" (UniqueName: \"kubernetes.io/projected/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-kube-api-access-bqnzf\") pod \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " Dec 08 15:04:02 crc kubenswrapper[4894]: I1208 15:04:02.992563 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-svc\") pod \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\" (UID: \"6a8d9046-faf2-413c-afc9-3c0eb3076b8e\") " Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.014396 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.031097 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79476f47f-8zstk"] Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.063432 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-kube-api-access-bqnzf" (OuterVolumeSpecName: "kube-api-access-bqnzf") pod "6a8d9046-faf2-413c-afc9-3c0eb3076b8e" (UID: "6a8d9046-faf2-413c-afc9-3c0eb3076b8e"). InnerVolumeSpecName "kube-api-access-bqnzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.094905 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f90688-7b17-4ec3-876f-fb59acc6586f-logs\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.094990 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69f90688-7b17-4ec3-876f-fb59acc6586f-horizon-secret-key\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.095075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-config-data\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.095106 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-scripts\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.095153 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl69m\" (UniqueName: \"kubernetes.io/projected/69f90688-7b17-4ec3-876f-fb59acc6586f-kube-api-access-zl69m\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.095196 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqnzf\" (UniqueName: \"kubernetes.io/projected/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-kube-api-access-bqnzf\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.130663 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a8d9046-faf2-413c-afc9-3c0eb3076b8e" (UID: "6a8d9046-faf2-413c-afc9-3c0eb3076b8e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.131445 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a8d9046-faf2-413c-afc9-3c0eb3076b8e" (UID: "6a8d9046-faf2-413c-afc9-3c0eb3076b8e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.146427 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a8d9046-faf2-413c-afc9-3c0eb3076b8e" (UID: "6a8d9046-faf2-413c-afc9-3c0eb3076b8e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.165194 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-config" (OuterVolumeSpecName: "config") pod "6a8d9046-faf2-413c-afc9-3c0eb3076b8e" (UID: "6a8d9046-faf2-413c-afc9-3c0eb3076b8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.167469 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a8d9046-faf2-413c-afc9-3c0eb3076b8e" (UID: "6a8d9046-faf2-413c-afc9-3c0eb3076b8e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.169856 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerStarted","Data":"1cc9664f3dcbe84bf0e0a37cc39b072eb13a4f8025b937080fbde8d36438b07b"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.177088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kzgzh" event={"ID":"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f","Type":"ContainerStarted","Data":"e366fbf43e4ab9f33b6bccddd40b102659722fc10c6fd28229bb934ee8d072e3"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.177139 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kzgzh" event={"ID":"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f","Type":"ContainerStarted","Data":"2130ce948bc6cbbd5163025152e5ebcf953cf54295111470dd8439c64e305824"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.190421 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c4bf45b45-vmwph" event={"ID":"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2","Type":"ContainerStarted","Data":"2e09635f053889c3bcaad95c49709371e46226fb02382c4572c0e71477fc90b7"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203092 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-scripts\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203161 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl69m\" (UniqueName: \"kubernetes.io/projected/69f90688-7b17-4ec3-876f-fb59acc6586f-kube-api-access-zl69m\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203195 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f90688-7b17-4ec3-876f-fb59acc6586f-logs\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203255 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69f90688-7b17-4ec3-876f-fb59acc6586f-horizon-secret-key\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203350 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-config-data\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203407 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203419 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203430 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203443 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.203454 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a8d9046-faf2-413c-afc9-3c0eb3076b8e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.204802 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-config-data\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.205538 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-scripts\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.206350 4894 generic.go:334] "Generic (PLEG): container finished" podID="6a8d9046-faf2-413c-afc9-3c0eb3076b8e" containerID="ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051" exitCode=0 Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.206418 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" event={"ID":"6a8d9046-faf2-413c-afc9-3c0eb3076b8e","Type":"ContainerDied","Data":"ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.206448 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" event={"ID":"6a8d9046-faf2-413c-afc9-3c0eb3076b8e","Type":"ContainerDied","Data":"ff2daef663aa9e837f3da0611539e04a23b462e001fee8a217023083e916317f"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.206468 4894 scope.go:117] "RemoveContainer" containerID="ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.206592 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-v2snv" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.209399 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f90688-7b17-4ec3-876f-fb59acc6586f-logs\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.211559 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kzgzh" podStartSLOduration=3.211542784 podStartE2EDuration="3.211542784s" podCreationTimestamp="2025-12-08 15:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:03.205789623 +0000 UTC m=+1064.305795738" watchObservedRunningTime="2025-12-08 15:04:03.211542784 +0000 UTC m=+1064.311548919" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.229758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69f90688-7b17-4ec3-876f-fb59acc6586f-horizon-secret-key\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.232382 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl69m\" (UniqueName: \"kubernetes.io/projected/69f90688-7b17-4ec3-876f-fb59acc6586f-kube-api-access-zl69m\") pod \"horizon-79476f47f-8zstk\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.333084 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.355255 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-749798c7fc-c62cr" event={"ID":"717a577a-7166-44b5-bc64-40553dfe66f8","Type":"ContainerStarted","Data":"f8e5ead62456b9875769da196ab01e8ff32e2ccd5d9f55d4fe3279cbd1c3cc4c"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.355306 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g4klr" event={"ID":"195700ca-d25a-470a-a6c4-73d4f4133b95","Type":"ContainerStarted","Data":"02412146543e671f6f92e070e4faa0ce4e09c3809a84e4eeef1659d66ef63d52"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.355323 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7b8gb" event={"ID":"85f57102-ceaf-484a-a0b3-c1ecf324339d","Type":"ContainerStarted","Data":"c3ab14a8234828cbf2885308e455f4dd432abb6a0b75092c3d53f3b35b7b5084"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.355337 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-htqq9" event={"ID":"bc5041f7-6cf0-46c8-8fc4-9616005d46b7","Type":"ContainerStarted","Data":"faa9d20bccbea00e7a42702e051752cab4d1e85396f3a5ed06a8ef428d55bb04"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.355349 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" event={"ID":"97e76d4c-be0b-4ed9-8e51-c5bed27d5514","Type":"ContainerStarted","Data":"f800e34eeb710ed4ea1b663212a1f495a517d4073fecce9547f742f6f70a162a"} Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.377913 4894 scope.go:117] "RemoveContainer" containerID="ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051" Dec 08 15:04:03 crc kubenswrapper[4894]: E1208 15:04:03.378753 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051\": container with ID starting with ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051 not found: ID does not exist" containerID="ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.378921 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051"} err="failed to get container status \"ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051\": rpc error: code = NotFound desc = could not find container \"ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051\": container with ID starting with ae27bb6f1c1b24ef397ac1aa5ed4a83edadbb0b1fa4ee5837862d25b88b0f051 not found: ID does not exist" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.436297 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-v2snv"] Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.474179 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-v2snv"] Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.476236 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-htqq9" podStartSLOduration=3.399127778 podStartE2EDuration="43.476214427s" podCreationTimestamp="2025-12-08 15:03:20 +0000 UTC" firstStartedPulling="2025-12-08 15:03:21.669581569 +0000 UTC m=+1022.769587674" lastFinishedPulling="2025-12-08 15:04:01.746668208 +0000 UTC m=+1062.846674323" observedRunningTime="2025-12-08 15:04:03.439663937 +0000 UTC m=+1064.539670052" watchObservedRunningTime="2025-12-08 15:04:03.476214427 +0000 UTC m=+1064.576220542" Dec 08 15:04:03 crc kubenswrapper[4894]: I1208 15:04:03.973533 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79476f47f-8zstk"] Dec 08 15:04:04 crc kubenswrapper[4894]: W1208 15:04:04.069036 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69f90688_7b17_4ec3_876f_fb59acc6586f.slice/crio-f7cb5a90a1ce05b62728aae9e78607e13741cb21504958e9a1363c01ef7fdaad WatchSource:0}: Error finding container f7cb5a90a1ce05b62728aae9e78607e13741cb21504958e9a1363c01ef7fdaad: Status 404 returned error can't find the container with id f7cb5a90a1ce05b62728aae9e78607e13741cb21504958e9a1363c01ef7fdaad Dec 08 15:04:04 crc kubenswrapper[4894]: E1208 15:04:04.249526 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739b42e0_365d_4eff_8724_d0de80c7d29f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice/crio-96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf51f62_aa82_474b_b3d4_4746043cf399.slice/crio-6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441\": RecentStats: unable to find data in memory cache]" Dec 08 15:04:04 crc kubenswrapper[4894]: I1208 15:04:04.300882 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79476f47f-8zstk" event={"ID":"69f90688-7b17-4ec3-876f-fb59acc6586f","Type":"ContainerStarted","Data":"f7cb5a90a1ce05b62728aae9e78607e13741cb21504958e9a1363c01ef7fdaad"} Dec 08 15:04:04 crc kubenswrapper[4894]: I1208 15:04:04.314658 4894 generic.go:334] "Generic (PLEG): container finished" podID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerID="043a6234a41046010426ee7943ab02197e6cf810b16af6279d11596187deee29" exitCode=0 Dec 08 15:04:04 crc kubenswrapper[4894]: I1208 15:04:04.314755 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" event={"ID":"97e76d4c-be0b-4ed9-8e51-c5bed27d5514","Type":"ContainerDied","Data":"043a6234a41046010426ee7943ab02197e6cf810b16af6279d11596187deee29"} Dec 08 15:04:04 crc kubenswrapper[4894]: I1208 15:04:04.314794 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" event={"ID":"97e76d4c-be0b-4ed9-8e51-c5bed27d5514","Type":"ContainerStarted","Data":"6e87f0ad99b65d8329594708817c1d0806b34a4d0a524d037510213144f1924c"} Dec 08 15:04:04 crc kubenswrapper[4894]: I1208 15:04:04.315120 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:04 crc kubenswrapper[4894]: I1208 15:04:04.349006 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" podStartSLOduration=4.348979982 podStartE2EDuration="4.348979982s" podCreationTimestamp="2025-12-08 15:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:04.34221061 +0000 UTC m=+1065.442216715" watchObservedRunningTime="2025-12-08 15:04:04.348979982 +0000 UTC m=+1065.448986107" Dec 08 15:04:05 crc kubenswrapper[4894]: I1208 15:04:05.246221 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8d9046-faf2-413c-afc9-3c0eb3076b8e" path="/var/lib/kubelet/pods/6a8d9046-faf2-413c-afc9-3c0eb3076b8e/volumes" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.744108 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c4bf45b45-vmwph"] Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.815525 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fb57bf44-rk2jn"] Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.833319 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fb57bf44-rk2jn"] Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.833502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.835629 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.905839 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79476f47f-8zstk"] Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.963253 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-797dcfbfb8-294k4"] Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.976375 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.977176 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-secret-key\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.977252 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzd27\" (UniqueName: \"kubernetes.io/projected/a44fcb75-8381-44cf-8c5b-2b1108232807-kube-api-access-qzd27\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.977494 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a44fcb75-8381-44cf-8c5b-2b1108232807-logs\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.977665 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-config-data\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.977776 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-tls-certs\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.977916 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-combined-ca-bundle\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:09 crc kubenswrapper[4894]: I1208 15:04:09.978019 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.023347 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797dcfbfb8-294k4"] Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083146 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-config-data\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083305 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9nm9\" (UniqueName: \"kubernetes.io/projected/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-kube-api-access-k9nm9\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083379 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-secret-key\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083409 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-horizon-tls-certs\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083444 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzd27\" (UniqueName: \"kubernetes.io/projected/a44fcb75-8381-44cf-8c5b-2b1108232807-kube-api-access-qzd27\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a44fcb75-8381-44cf-8c5b-2b1108232807-logs\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083512 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-scripts\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083551 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-combined-ca-bundle\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083587 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-config-data\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083611 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-tls-certs\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083644 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-horizon-secret-key\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083686 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-combined-ca-bundle\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083713 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-logs\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.083750 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.084709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.087802 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a44fcb75-8381-44cf-8c5b-2b1108232807-logs\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.088482 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-config-data\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.099288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-combined-ca-bundle\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.099324 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-secret-key\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.116702 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-tls-certs\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.140030 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzd27\" (UniqueName: \"kubernetes.io/projected/a44fcb75-8381-44cf-8c5b-2b1108232807-kube-api-access-qzd27\") pod \"horizon-7fb57bf44-rk2jn\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.168604 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.184942 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-horizon-tls-certs\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.185033 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-scripts\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.185080 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-combined-ca-bundle\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.185132 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-horizon-secret-key\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.185177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-logs\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.185240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-config-data\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.185268 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9nm9\" (UniqueName: \"kubernetes.io/projected/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-kube-api-access-k9nm9\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.188639 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-scripts\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.188976 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-logs\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.190047 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-config-data\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.203488 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-horizon-tls-certs\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.204073 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-horizon-secret-key\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.207852 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-combined-ca-bundle\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.218158 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9nm9\" (UniqueName: \"kubernetes.io/projected/c3009b99-5e3b-4f39-b62f-0e04aeb371f2-kube-api-access-k9nm9\") pod \"horizon-797dcfbfb8-294k4\" (UID: \"c3009b99-5e3b-4f39-b62f-0e04aeb371f2\") " pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:10 crc kubenswrapper[4894]: I1208 15:04:10.299179 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:11 crc kubenswrapper[4894]: I1208 15:04:11.396980 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:11 crc kubenswrapper[4894]: I1208 15:04:11.434216 4894 generic.go:334] "Generic (PLEG): container finished" podID="b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" containerID="548607c0f720255da6e0f5953d7edf31caf09aea918dc6d6c6ddd0e5569d351a" exitCode=0 Dec 08 15:04:11 crc kubenswrapper[4894]: I1208 15:04:11.434258 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5tvxs" event={"ID":"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29","Type":"ContainerDied","Data":"548607c0f720255da6e0f5953d7edf31caf09aea918dc6d6c6ddd0e5569d351a"} Dec 08 15:04:11 crc kubenswrapper[4894]: I1208 15:04:11.480158 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-mfp9l"] Dec 08 15:04:11 crc kubenswrapper[4894]: I1208 15:04:11.480442 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="dnsmasq-dns" containerID="cri-o://dcb2a8fdc65b7cad7f2ea3d18f923d001adad65a1e124f36950210a9d83159e7" gracePeriod=10 Dec 08 15:04:12 crc kubenswrapper[4894]: I1208 15:04:12.451309 4894 generic.go:334] "Generic (PLEG): container finished" podID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerID="dcb2a8fdc65b7cad7f2ea3d18f923d001adad65a1e124f36950210a9d83159e7" exitCode=0 Dec 08 15:04:12 crc kubenswrapper[4894]: I1208 15:04:12.451386 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" event={"ID":"b5f234d1-2387-40b5-bfc9-0b899c6e7b07","Type":"ContainerDied","Data":"dcb2a8fdc65b7cad7f2ea3d18f923d001adad65a1e124f36950210a9d83159e7"} Dec 08 15:04:14 crc kubenswrapper[4894]: E1208 15:04:14.492897 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739b42e0_365d_4eff_8724_d0de80c7d29f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb83cfcb_cfb9_4c84_919a_f4408f4962c5.slice/crio-96e9d99b5db82835246f0d22c49a5c7a09e166980e53ed132cb211abcfb1b3f9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf51f62_aa82_474b_b3d4_4746043cf399.slice/crio-6d0327c0bb3258571af020752b0849b2547a6b63ba1228983cf215e04c5a5441\": RecentStats: unable to find data in memory cache]" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.743549 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.744532 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n654h544h4h549h54bh648h549h64dhbh5bdh64h9ch7bh5cch657h5bch58fh58fhbh59ch57fh65bh94hf9h75h64dh6ch9bhb7h578hch58bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqmmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-c4bf45b45-vmwph_openstack(586511a0-9e32-4c9b-bbff-1cb74f0a0cc2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.748595 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-c4bf45b45-vmwph" podUID="586511a0-9e32-4c9b-bbff-1cb74f0a0cc2" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.755476 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.755625 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd4h58h95hfbh66h5cfh687h64h5fdh568h5ddhd5h597h8dh54h5c5hc7hbh559h5b8h557h56dhb4h5b9h5fch6dh654hd6hfch666h559h5bfq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qqnj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-749798c7fc-c62cr_openstack(717a577a-7166-44b5-bc64-40553dfe66f8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.757594 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-749798c7fc-c62cr" podUID="717a577a-7166-44b5-bc64-40553dfe66f8" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.767050 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.767223 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n579hf9h5b8h64hcfh679h76h685h74h5bh597h654hcbh5cchchf7h5f9h598hd9h5c8h644h575h5c6h5bfh589h676h56ch655hdfh79h695hfq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zl69m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-79476f47f-8zstk_openstack(69f90688-7b17-4ec3-876f-fb59acc6586f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:04:17 crc kubenswrapper[4894]: E1208 15:04:17.769965 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-79476f47f-8zstk" podUID="69f90688-7b17-4ec3-876f-fb59acc6586f" Dec 08 15:04:19 crc kubenswrapper[4894]: I1208 15:04:19.397307 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 08 15:04:22 crc kubenswrapper[4894]: I1208 15:04:22.543968 4894 generic.go:334] "Generic (PLEG): container finished" podID="bc5041f7-6cf0-46c8-8fc4-9616005d46b7" containerID="faa9d20bccbea00e7a42702e051752cab4d1e85396f3a5ed06a8ef428d55bb04" exitCode=0 Dec 08 15:04:22 crc kubenswrapper[4894]: I1208 15:04:22.544016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-htqq9" event={"ID":"bc5041f7-6cf0-46c8-8fc4-9616005d46b7","Type":"ContainerDied","Data":"faa9d20bccbea00e7a42702e051752cab4d1e85396f3a5ed06a8ef428d55bb04"} Dec 08 15:04:24 crc kubenswrapper[4894]: I1208 15:04:24.398974 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 08 15:04:28 crc kubenswrapper[4894]: I1208 15:04:28.334537 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:04:28 crc kubenswrapper[4894]: I1208 15:04:28.339775 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:28 crc kubenswrapper[4894]: I1208 15:04:28.356119 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:28 crc kubenswrapper[4894]: I1208 15:04:28.371646 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:28 crc kubenswrapper[4894]: I1208 15:04:28.398630 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503122 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717a577a-7166-44b5-bc64-40553dfe66f8-logs\") pod \"717a577a-7166-44b5-bc64-40553dfe66f8\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503199 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-config-data\") pod \"717a577a-7166-44b5-bc64-40553dfe66f8\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503264 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-swift-storage-0\") pod \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503289 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqnj2\" (UniqueName: \"kubernetes.io/projected/717a577a-7166-44b5-bc64-40553dfe66f8-kube-api-access-qqnj2\") pod \"717a577a-7166-44b5-bc64-40553dfe66f8\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503343 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-config\") pod \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503385 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-config-data\") pod \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503494 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/717a577a-7166-44b5-bc64-40553dfe66f8-logs" (OuterVolumeSpecName: "logs") pod "717a577a-7166-44b5-bc64-40553dfe66f8" (UID: "717a577a-7166-44b5-bc64-40553dfe66f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503512 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4sxm\" (UniqueName: \"kubernetes.io/projected/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-kube-api-access-b4sxm\") pod \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503550 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-sb\") pod \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503570 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-config-data\") pod \"69f90688-7b17-4ec3-876f-fb59acc6586f\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503604 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-horizon-secret-key\") pod \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503656 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/717a577a-7166-44b5-bc64-40553dfe66f8-horizon-secret-key\") pod \"717a577a-7166-44b5-bc64-40553dfe66f8\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503696 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-svc\") pod \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503721 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-combined-ca-bundle\") pod \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503741 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-scripts\") pod \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503762 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-scripts\") pod \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-fernet-keys\") pod \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503843 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl69m\" (UniqueName: \"kubernetes.io/projected/69f90688-7b17-4ec3-876f-fb59acc6586f-kube-api-access-zl69m\") pod \"69f90688-7b17-4ec3-876f-fb59acc6586f\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503872 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f90688-7b17-4ec3-876f-fb59acc6586f-logs\") pod \"69f90688-7b17-4ec3-876f-fb59acc6586f\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503908 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-logs\") pod \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503952 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-nb\") pod \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503970 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-scripts\") pod \"69f90688-7b17-4ec3-876f-fb59acc6586f\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.503996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-config-data\") pod \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504017 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqmmq\" (UniqueName: \"kubernetes.io/projected/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-kube-api-access-nqmmq\") pod \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\" (UID: \"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504048 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc6lt\" (UniqueName: \"kubernetes.io/projected/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-kube-api-access-fc6lt\") pod \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\" (UID: \"b5f234d1-2387-40b5-bfc9-0b899c6e7b07\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504111 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-credential-keys\") pod \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\" (UID: \"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504152 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-scripts\") pod \"717a577a-7166-44b5-bc64-40553dfe66f8\" (UID: \"717a577a-7166-44b5-bc64-40553dfe66f8\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504181 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69f90688-7b17-4ec3-876f-fb59acc6586f-horizon-secret-key\") pod \"69f90688-7b17-4ec3-876f-fb59acc6586f\" (UID: \"69f90688-7b17-4ec3-876f-fb59acc6586f\") " Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504392 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-config-data" (OuterVolumeSpecName: "config-data") pod "69f90688-7b17-4ec3-876f-fb59acc6586f" (UID: "69f90688-7b17-4ec3-876f-fb59acc6586f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504781 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/717a577a-7166-44b5-bc64-40553dfe66f8-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.504809 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.505653 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69f90688-7b17-4ec3-876f-fb59acc6586f-logs" (OuterVolumeSpecName: "logs") pod "69f90688-7b17-4ec3-876f-fb59acc6586f" (UID: "69f90688-7b17-4ec3-876f-fb59acc6586f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.506058 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-config-data" (OuterVolumeSpecName: "config-data") pod "717a577a-7166-44b5-bc64-40553dfe66f8" (UID: "717a577a-7166-44b5-bc64-40553dfe66f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.507237 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-scripts" (OuterVolumeSpecName: "scripts") pod "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2" (UID: "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.509537 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-config-data" (OuterVolumeSpecName: "config-data") pod "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2" (UID: "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.510058 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-scripts" (OuterVolumeSpecName: "scripts") pod "717a577a-7166-44b5-bc64-40553dfe66f8" (UID: "717a577a-7166-44b5-bc64-40553dfe66f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.510167 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-logs" (OuterVolumeSpecName: "logs") pod "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2" (UID: "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.511772 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-kube-api-access-b4sxm" (OuterVolumeSpecName: "kube-api-access-b4sxm") pod "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" (UID: "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29"). InnerVolumeSpecName "kube-api-access-b4sxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.512298 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2" (UID: "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.516664 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69f90688-7b17-4ec3-876f-fb59acc6586f-kube-api-access-zl69m" (OuterVolumeSpecName: "kube-api-access-zl69m") pod "69f90688-7b17-4ec3-876f-fb59acc6586f" (UID: "69f90688-7b17-4ec3-876f-fb59acc6586f"). InnerVolumeSpecName "kube-api-access-zl69m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.517132 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" (UID: "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.517318 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717a577a-7166-44b5-bc64-40553dfe66f8-kube-api-access-qqnj2" (OuterVolumeSpecName: "kube-api-access-qqnj2") pod "717a577a-7166-44b5-bc64-40553dfe66f8" (UID: "717a577a-7166-44b5-bc64-40553dfe66f8"). InnerVolumeSpecName "kube-api-access-qqnj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.517904 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-scripts" (OuterVolumeSpecName: "scripts") pod "69f90688-7b17-4ec3-876f-fb59acc6586f" (UID: "69f90688-7b17-4ec3-876f-fb59acc6586f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.520271 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-kube-api-access-fc6lt" (OuterVolumeSpecName: "kube-api-access-fc6lt") pod "b5f234d1-2387-40b5-bfc9-0b899c6e7b07" (UID: "b5f234d1-2387-40b5-bfc9-0b899c6e7b07"). InnerVolumeSpecName "kube-api-access-fc6lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.520764 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-kube-api-access-nqmmq" (OuterVolumeSpecName: "kube-api-access-nqmmq") pod "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2" (UID: "586511a0-9e32-4c9b-bbff-1cb74f0a0cc2"). InnerVolumeSpecName "kube-api-access-nqmmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.522530 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69f90688-7b17-4ec3-876f-fb59acc6586f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "69f90688-7b17-4ec3-876f-fb59acc6586f" (UID: "69f90688-7b17-4ec3-876f-fb59acc6586f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.528062 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" (UID: "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.528577 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717a577a-7166-44b5-bc64-40553dfe66f8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "717a577a-7166-44b5-bc64-40553dfe66f8" (UID: "717a577a-7166-44b5-bc64-40553dfe66f8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.529484 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-scripts" (OuterVolumeSpecName: "scripts") pod "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" (UID: "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.569115 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-config-data" (OuterVolumeSpecName: "config-data") pod "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" (UID: "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.573922 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-config" (OuterVolumeSpecName: "config") pod "b5f234d1-2387-40b5-bfc9-0b899c6e7b07" (UID: "b5f234d1-2387-40b5-bfc9-0b899c6e7b07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.579794 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" (UID: "b9a80eb2-a9be-4f4f-84c0-8fd96392cc29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.595991 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5f234d1-2387-40b5-bfc9-0b899c6e7b07" (UID: "b5f234d1-2387-40b5-bfc9-0b899c6e7b07"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.598998 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c4bf45b45-vmwph" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.599039 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c4bf45b45-vmwph" event={"ID":"586511a0-9e32-4c9b-bbff-1cb74f0a0cc2","Type":"ContainerDied","Data":"2e09635f053889c3bcaad95c49709371e46226fb02382c4572c0e71477fc90b7"} Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.601209 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-749798c7fc-c62cr" event={"ID":"717a577a-7166-44b5-bc64-40553dfe66f8","Type":"ContainerDied","Data":"f8e5ead62456b9875769da196ab01e8ff32e2ccd5d9f55d4fe3279cbd1c3cc4c"} Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.601299 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749798c7fc-c62cr" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607168 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69f90688-7b17-4ec3-876f-fb59acc6586f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607193 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607202 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607212 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqnj2\" (UniqueName: \"kubernetes.io/projected/717a577a-7166-44b5-bc64-40553dfe66f8-kube-api-access-qqnj2\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607221 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607230 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607239 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4sxm\" (UniqueName: \"kubernetes.io/projected/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-kube-api-access-b4sxm\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607248 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607258 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/717a577a-7166-44b5-bc64-40553dfe66f8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607268 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607276 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607284 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607292 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607300 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl69m\" (UniqueName: \"kubernetes.io/projected/69f90688-7b17-4ec3-876f-fb59acc6586f-kube-api-access-zl69m\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607308 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f90688-7b17-4ec3-876f-fb59acc6586f-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607316 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607323 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69f90688-7b17-4ec3-876f-fb59acc6586f-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607333 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607342 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqmmq\" (UniqueName: \"kubernetes.io/projected/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2-kube-api-access-nqmmq\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607351 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc6lt\" (UniqueName: \"kubernetes.io/projected/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-kube-api-access-fc6lt\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607359 4894 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.607367 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/717a577a-7166-44b5-bc64-40553dfe66f8-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.608465 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5tvxs" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.608707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5tvxs" event={"ID":"b9a80eb2-a9be-4f4f-84c0-8fd96392cc29","Type":"ContainerDied","Data":"4e0281815a25957438232fc2a93f474087b4e825db050d7d325002f9613e124b"} Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.608745 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e0281815a25957438232fc2a93f474087b4e825db050d7d325002f9613e124b" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.613530 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5f234d1-2387-40b5-bfc9-0b899c6e7b07" (UID: "b5f234d1-2387-40b5-bfc9-0b899c6e7b07"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.614359 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5f234d1-2387-40b5-bfc9-0b899c6e7b07" (UID: "b5f234d1-2387-40b5-bfc9-0b899c6e7b07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.615172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" event={"ID":"b5f234d1-2387-40b5-bfc9-0b899c6e7b07","Type":"ContainerDied","Data":"79a7bee0cf82c48542552adf4a1bd327b4a6493396fbaf9a020323b1300b7e0c"} Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.615220 4894 scope.go:117] "RemoveContainer" containerID="dcb2a8fdc65b7cad7f2ea3d18f923d001adad65a1e124f36950210a9d83159e7" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.615237 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.618303 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79476f47f-8zstk" event={"ID":"69f90688-7b17-4ec3-876f-fb59acc6586f","Type":"ContainerDied","Data":"f7cb5a90a1ce05b62728aae9e78607e13741cb21504958e9a1363c01ef7fdaad"} Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.618596 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79476f47f-8zstk" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.625572 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5f234d1-2387-40b5-bfc9-0b899c6e7b07" (UID: "b5f234d1-2387-40b5-bfc9-0b899c6e7b07"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.710122 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.710169 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.710179 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5f234d1-2387-40b5-bfc9-0b899c6e7b07-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.737170 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fb57bf44-rk2jn"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.848994 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-749798c7fc-c62cr"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.882133 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-749798c7fc-c62cr"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.938938 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c4bf45b45-vmwph"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.949188 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c4bf45b45-vmwph"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.974478 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79476f47f-8zstk"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.980032 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79476f47f-8zstk"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.991237 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-mfp9l"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:28.994563 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-mfp9l"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.215996 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="586511a0-9e32-4c9b-bbff-1cb74f0a0cc2" path="/var/lib/kubelet/pods/586511a0-9e32-4c9b-bbff-1cb74f0a0cc2/volumes" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.216956 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69f90688-7b17-4ec3-876f-fb59acc6586f" path="/var/lib/kubelet/pods/69f90688-7b17-4ec3-876f-fb59acc6586f/volumes" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.217713 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="717a577a-7166-44b5-bc64-40553dfe66f8" path="/var/lib/kubelet/pods/717a577a-7166-44b5-bc64-40553dfe66f8/volumes" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.218380 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" path="/var/lib/kubelet/pods/b5f234d1-2387-40b5-bfc9-0b899c6e7b07/volumes" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.399547 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-mfp9l" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.475332 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5tvxs"] Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.481501 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5tvxs"] Dec 08 15:04:29 crc kubenswrapper[4894]: E1208 15:04:29.497326 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 08 15:04:29 crc kubenswrapper[4894]: E1208 15:04:29.497537 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5z2l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-7b8gb_openstack(85f57102-ceaf-484a-a0b3-c1ecf324339d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:04:29 crc kubenswrapper[4894]: E1208 15:04:29.498894 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-7b8gb" podUID="85f57102-ceaf-484a-a0b3-c1ecf324339d" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.579157 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-c642j"] Dec 08 15:04:29 crc kubenswrapper[4894]: E1208 15:04:29.579735 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="dnsmasq-dns" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.579761 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="dnsmasq-dns" Dec 08 15:04:29 crc kubenswrapper[4894]: E1208 15:04:29.579783 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="init" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.579794 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="init" Dec 08 15:04:29 crc kubenswrapper[4894]: E1208 15:04:29.579833 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" containerName="keystone-bootstrap" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.579845 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" containerName="keystone-bootstrap" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.580085 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5f234d1-2387-40b5-bfc9-0b899c6e7b07" containerName="dnsmasq-dns" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.580126 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" containerName="keystone-bootstrap" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.581041 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.589267 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.589284 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.589411 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.589654 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dscs7" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.589785 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.603234 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c642j"] Dec 08 15:04:29 crc kubenswrapper[4894]: E1208 15:04:29.639167 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-7b8gb" podUID="85f57102-ceaf-484a-a0b3-c1ecf324339d" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.662495 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-scripts\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.662549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-combined-ca-bundle\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.662592 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-fernet-keys\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.662657 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-config-data\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.662680 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-credential-keys\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.662790 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76qcv\" (UniqueName: \"kubernetes.io/projected/77a9427b-9569-447d-aa60-e3be638e2d4c-kube-api-access-76qcv\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.765968 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-config-data\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.766355 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-credential-keys\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.766571 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76qcv\" (UniqueName: \"kubernetes.io/projected/77a9427b-9569-447d-aa60-e3be638e2d4c-kube-api-access-76qcv\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.766733 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-scripts\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.766961 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-combined-ca-bundle\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.767067 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-fernet-keys\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.776008 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-config-data\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.777354 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-fernet-keys\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.781351 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-credential-keys\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.783454 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-combined-ca-bundle\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.789603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76qcv\" (UniqueName: \"kubernetes.io/projected/77a9427b-9569-447d-aa60-e3be638e2d4c-kube-api-access-76qcv\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.794305 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-scripts\") pod \"keystone-bootstrap-c642j\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:29 crc kubenswrapper[4894]: I1208 15:04:29.931337 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:31 crc kubenswrapper[4894]: W1208 15:04:31.030549 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda44fcb75_8381_44cf_8c5b_2b1108232807.slice/crio-be906960d703b7c0096a82357a5f2d9e67c1c8a1579ff7d541bb3e01f93bc5ff WatchSource:0}: Error finding container be906960d703b7c0096a82357a5f2d9e67c1c8a1579ff7d541bb3e01f93bc5ff: Status 404 returned error can't find the container with id be906960d703b7c0096a82357a5f2d9e67c1c8a1579ff7d541bb3e01f93bc5ff Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.048883 4894 scope.go:117] "RemoveContainer" containerID="e68896aff1f151fb7ab4bf41adff909ce5419d7da4b625b06f94fcd93933b9aa" Dec 08 15:04:31 crc kubenswrapper[4894]: E1208 15:04:31.095308 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 08 15:04:31 crc kubenswrapper[4894]: E1208 15:04:31.095491 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9hbvt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7nxjg_openstack(06177501-5baa-49e0-94bd-157c6f079c05): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:04:31 crc kubenswrapper[4894]: E1208 15:04:31.097560 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7nxjg" podUID="06177501-5baa-49e0-94bd-157c6f079c05" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.235092 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a80eb2-a9be-4f4f-84c0-8fd96392cc29" path="/var/lib/kubelet/pods/b9a80eb2-a9be-4f4f-84c0-8fd96392cc29/volumes" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.339440 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-htqq9" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.407510 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-combined-ca-bundle\") pod \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.407767 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-db-sync-config-data\") pod \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.407949 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-config-data\") pod \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.408049 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r6tt\" (UniqueName: \"kubernetes.io/projected/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-kube-api-access-7r6tt\") pod \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\" (UID: \"bc5041f7-6cf0-46c8-8fc4-9616005d46b7\") " Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.414962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bc5041f7-6cf0-46c8-8fc4-9616005d46b7" (UID: "bc5041f7-6cf0-46c8-8fc4-9616005d46b7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.414974 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-kube-api-access-7r6tt" (OuterVolumeSpecName: "kube-api-access-7r6tt") pod "bc5041f7-6cf0-46c8-8fc4-9616005d46b7" (UID: "bc5041f7-6cf0-46c8-8fc4-9616005d46b7"). InnerVolumeSpecName "kube-api-access-7r6tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.451633 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc5041f7-6cf0-46c8-8fc4-9616005d46b7" (UID: "bc5041f7-6cf0-46c8-8fc4-9616005d46b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.497850 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-config-data" (OuterVolumeSpecName: "config-data") pod "bc5041f7-6cf0-46c8-8fc4-9616005d46b7" (UID: "bc5041f7-6cf0-46c8-8fc4-9616005d46b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.500999 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797dcfbfb8-294k4"] Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.515417 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.515473 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.515490 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r6tt\" (UniqueName: \"kubernetes.io/projected/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-kube-api-access-7r6tt\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.515507 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc5041f7-6cf0-46c8-8fc4-9616005d46b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.565190 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-c642j"] Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.761551 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797dcfbfb8-294k4" event={"ID":"c3009b99-5e3b-4f39-b62f-0e04aeb371f2","Type":"ContainerStarted","Data":"fc3daa1a24b666178d279e57ed627bf7f351ed1916c691f2200b97ba3455e7c7"} Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.764177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g4klr" event={"ID":"195700ca-d25a-470a-a6c4-73d4f4133b95","Type":"ContainerStarted","Data":"90aebf031ec3f48a436b2122f40dd850b0eeee78920dc26741dd010da6c2f482"} Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.766323 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c642j" event={"ID":"77a9427b-9569-447d-aa60-e3be638e2d4c","Type":"ContainerStarted","Data":"e718982841f24ff257045e10dc0fd07d57cc3c68d60eb78e5a394d703d3fa7c3"} Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.770180 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-htqq9" event={"ID":"bc5041f7-6cf0-46c8-8fc4-9616005d46b7","Type":"ContainerDied","Data":"e006cf5770d302bc7de70ae01dc3e2c9dbe12a53a58038dcf8a679dfb3544e7d"} Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.770215 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e006cf5770d302bc7de70ae01dc3e2c9dbe12a53a58038dcf8a679dfb3544e7d" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.770303 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-htqq9" Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.780480 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb57bf44-rk2jn" event={"ID":"a44fcb75-8381-44cf-8c5b-2b1108232807","Type":"ContainerStarted","Data":"be906960d703b7c0096a82357a5f2d9e67c1c8a1579ff7d541bb3e01f93bc5ff"} Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.785366 4894 generic.go:334] "Generic (PLEG): container finished" podID="bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" containerID="e366fbf43e4ab9f33b6bccddd40b102659722fc10c6fd28229bb934ee8d072e3" exitCode=0 Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.785503 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kzgzh" event={"ID":"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f","Type":"ContainerDied","Data":"e366fbf43e4ab9f33b6bccddd40b102659722fc10c6fd28229bb934ee8d072e3"} Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.791336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerStarted","Data":"a258a59f2c327d7149b40ebefb52a70112b73cda3bd4d7b858fc52d4154f5ee8"} Dec 08 15:04:31 crc kubenswrapper[4894]: I1208 15:04:31.794311 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-g4klr" podStartSLOduration=6.415746246 podStartE2EDuration="31.794289818s" podCreationTimestamp="2025-12-08 15:04:00 +0000 UTC" firstStartedPulling="2025-12-08 15:04:02.810077659 +0000 UTC m=+1063.910083774" lastFinishedPulling="2025-12-08 15:04:28.188621231 +0000 UTC m=+1089.288627346" observedRunningTime="2025-12-08 15:04:31.785599815 +0000 UTC m=+1092.885605930" watchObservedRunningTime="2025-12-08 15:04:31.794289818 +0000 UTC m=+1092.894295933" Dec 08 15:04:31 crc kubenswrapper[4894]: E1208 15:04:31.797377 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7nxjg" podUID="06177501-5baa-49e0-94bd-157c6f079c05" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.824584 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797dcfbfb8-294k4" event={"ID":"c3009b99-5e3b-4f39-b62f-0e04aeb371f2","Type":"ContainerStarted","Data":"26aa8cb1d14ccec057ce66321d96a4ae87959b47df06b5e4dae0cbc9174fbc83"} Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.825161 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797dcfbfb8-294k4" event={"ID":"c3009b99-5e3b-4f39-b62f-0e04aeb371f2","Type":"ContainerStarted","Data":"8676983ce55c13d18938fc3ba3d59c21760af1d33d2925af1cd1cb43b5cbffae"} Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.827053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c642j" event={"ID":"77a9427b-9569-447d-aa60-e3be638e2d4c","Type":"ContainerStarted","Data":"d9a002ecf2a43b769b650f5eaf45af28c9c5bcf093c16c5beffca14bd9a79304"} Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.841258 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb57bf44-rk2jn" event={"ID":"a44fcb75-8381-44cf-8c5b-2b1108232807","Type":"ContainerStarted","Data":"31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844"} Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.841366 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb57bf44-rk2jn" event={"ID":"a44fcb75-8381-44cf-8c5b-2b1108232807","Type":"ContainerStarted","Data":"49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7"} Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.851668 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-797dcfbfb8-294k4" podStartSLOduration=23.851652869 podStartE2EDuration="23.851652869s" podCreationTimestamp="2025-12-08 15:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:32.850351009 +0000 UTC m=+1093.950357124" watchObservedRunningTime="2025-12-08 15:04:32.851652869 +0000 UTC m=+1093.951658984" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.896921 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-hz8tj"] Dec 08 15:04:32 crc kubenswrapper[4894]: E1208 15:04:32.897477 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc5041f7-6cf0-46c8-8fc4-9616005d46b7" containerName="glance-db-sync" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.897504 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc5041f7-6cf0-46c8-8fc4-9616005d46b7" containerName="glance-db-sync" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.897750 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc5041f7-6cf0-46c8-8fc4-9616005d46b7" containerName="glance-db-sync" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.898164 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-c642j" podStartSLOduration=3.898151482 podStartE2EDuration="3.898151482s" podCreationTimestamp="2025-12-08 15:04:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:32.883686417 +0000 UTC m=+1093.983692532" watchObservedRunningTime="2025-12-08 15:04:32.898151482 +0000 UTC m=+1093.998157597" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.899165 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.969582 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.969643 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.969741 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-config\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.969759 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt7fr\" (UniqueName: \"kubernetes.io/projected/6a32cc27-948b-47d6-bd40-26360911278c-kube-api-access-vt7fr\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.969881 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:32 crc kubenswrapper[4894]: I1208 15:04:32.969899 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.000951 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-hz8tj"] Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.001430 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7fb57bf44-rk2jn" podStartSLOduration=23.58931365 podStartE2EDuration="24.001393269s" podCreationTimestamp="2025-12-08 15:04:09 +0000 UTC" firstStartedPulling="2025-12-08 15:04:31.049880759 +0000 UTC m=+1092.149886874" lastFinishedPulling="2025-12-08 15:04:31.461960378 +0000 UTC m=+1092.561966493" observedRunningTime="2025-12-08 15:04:32.91845302 +0000 UTC m=+1094.018459155" watchObservedRunningTime="2025-12-08 15:04:33.001393269 +0000 UTC m=+1094.101399384" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.072375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.072440 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.072525 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-config\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.072545 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt7fr\" (UniqueName: \"kubernetes.io/projected/6a32cc27-948b-47d6-bd40-26360911278c-kube-api-access-vt7fr\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.072603 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.072628 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.074119 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.074125 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-config\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.074762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.075177 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.076900 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.119851 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt7fr\" (UniqueName: \"kubernetes.io/projected/6a32cc27-948b-47d6-bd40-26360911278c-kube-api-access-vt7fr\") pod \"dnsmasq-dns-8b5c85b87-hz8tj\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.252771 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.740131 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.753596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.763566 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.763915 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.775318 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lg9b9" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.817070 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.820349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.820504 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-config-data\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.820551 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-logs\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.820627 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-scripts\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.820679 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.820723 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.820779 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcr6d\" (UniqueName: \"kubernetes.io/projected/d612a58d-1418-42bb-b4d2-490ec2675699-kube-api-access-hcr6d\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.922451 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-config-data\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.923796 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-logs\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.924004 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-scripts\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.924125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.924210 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.924304 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcr6d\" (UniqueName: \"kubernetes.io/projected/d612a58d-1418-42bb-b4d2-490ec2675699-kube-api-access-hcr6d\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.924616 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.928639 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.928789 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-logs\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.929118 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.957758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-scripts\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.959704 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-config-data\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.961220 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcr6d\" (UniqueName: \"kubernetes.io/projected/d612a58d-1418-42bb-b4d2-490ec2675699-kube-api-access-hcr6d\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.966043 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:33 crc kubenswrapper[4894]: I1208 15:04:33.969373 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.044646 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.076892 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.088455 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:34 crc kubenswrapper[4894]: E1208 15:04:34.088904 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" containerName="neutron-db-sync" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.088923 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" containerName="neutron-db-sync" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.090592 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" containerName="neutron-db-sync" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.112155 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.112272 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.116101 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232230 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-combined-ca-bundle\") pod \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232429 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75k8k\" (UniqueName: \"kubernetes.io/projected/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-kube-api-access-75k8k\") pod \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232446 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-config\") pod \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\" (UID: \"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f\") " Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232679 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232715 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232764 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs2bv\" (UniqueName: \"kubernetes.io/projected/5689da95-4134-41d1-9ab5-9c18f1995438-kube-api-access-zs2bv\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232787 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232861 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232894 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.232945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-logs\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.240103 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-kube-api-access-75k8k" (OuterVolumeSpecName: "kube-api-access-75k8k") pod "bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" (UID: "bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f"). InnerVolumeSpecName "kube-api-access-75k8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.281010 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-config" (OuterVolumeSpecName: "config") pod "bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" (UID: "bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.290745 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" (UID: "bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.336690 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.336996 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-logs\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338212 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338297 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338393 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs2bv\" (UniqueName: \"kubernetes.io/projected/5689da95-4134-41d1-9ab5-9c18f1995438-kube-api-access-zs2bv\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338469 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338714 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338772 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75k8k\" (UniqueName: \"kubernetes.io/projected/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-kube-api-access-75k8k\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.338844 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.339160 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.339230 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-logs\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.340187 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.352984 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.354293 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.372166 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs2bv\" (UniqueName: \"kubernetes.io/projected/5689da95-4134-41d1-9ab5-9c18f1995438-kube-api-access-zs2bv\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.387547 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-hz8tj"] Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.389599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.393663 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.464706 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.871586 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.875403 4894 generic.go:334] "Generic (PLEG): container finished" podID="6a32cc27-948b-47d6-bd40-26360911278c" containerID="d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c" exitCode=0 Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.875509 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" event={"ID":"6a32cc27-948b-47d6-bd40-26360911278c","Type":"ContainerDied","Data":"d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c"} Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.875561 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" event={"ID":"6a32cc27-948b-47d6-bd40-26360911278c","Type":"ContainerStarted","Data":"314305dea69d85b069af39bf7fda54cef35d5448b75c70f35bed00aeaaae632c"} Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.879400 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kzgzh" event={"ID":"bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f","Type":"ContainerDied","Data":"2130ce948bc6cbbd5163025152e5ebcf953cf54295111470dd8439c64e305824"} Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.879463 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2130ce948bc6cbbd5163025152e5ebcf953cf54295111470dd8439c64e305824" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.879515 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kzgzh" Dec 08 15:04:34 crc kubenswrapper[4894]: I1208 15:04:34.887282 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerStarted","Data":"1af977c38788e3c71dbb4eee715a307b85476868c54ad8969918772446207d0e"} Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.110971 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.330107 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-hz8tj"] Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.384594 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-8sqvg"] Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.388792 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.397197 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-8sqvg"] Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.479175 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpg6h\" (UniqueName: \"kubernetes.io/projected/8be96a69-9240-4484-96d0-8b5eb9e44ead-kube-api-access-gpg6h\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.479261 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.479342 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-config\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.479377 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.479480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.479549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.547895 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6889cd597d-q9lvm"] Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.549873 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.553272 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xr6bk" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.559338 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.559636 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.559770 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.581053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.581099 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.581163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpg6h\" (UniqueName: \"kubernetes.io/projected/8be96a69-9240-4484-96d0-8b5eb9e44ead-kube-api-access-gpg6h\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.581192 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.581236 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-config\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.581269 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.582333 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.582890 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.583511 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.583681 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.584430 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-config\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.611890 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6889cd597d-q9lvm"] Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.646208 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpg6h\" (UniqueName: \"kubernetes.io/projected/8be96a69-9240-4484-96d0-8b5eb9e44ead-kube-api-access-gpg6h\") pod \"dnsmasq-dns-84b966f6c9-8sqvg\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.682486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-ovndb-tls-certs\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.682718 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-combined-ca-bundle\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.683413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-config\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.683606 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqgp7\" (UniqueName: \"kubernetes.io/projected/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-kube-api-access-rqgp7\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.683697 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-httpd-config\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.749514 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.786913 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-ovndb-tls-certs\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.787271 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-combined-ca-bundle\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.787485 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-config\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.787680 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqgp7\" (UniqueName: \"kubernetes.io/projected/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-kube-api-access-rqgp7\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.787787 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-httpd-config\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.795399 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-httpd-config\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.800552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-combined-ca-bundle\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.801239 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-config\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.828015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqgp7\" (UniqueName: \"kubernetes.io/projected/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-kube-api-access-rqgp7\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.851761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-ovndb-tls-certs\") pod \"neutron-6889cd597d-q9lvm\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.887659 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.911650 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5689da95-4134-41d1-9ab5-9c18f1995438","Type":"ContainerStarted","Data":"e88539449f678d9e373346cb6c973a0d822cbdeefd5ed38f38298a354ea62e75"} Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.924307 4894 generic.go:334] "Generic (PLEG): container finished" podID="195700ca-d25a-470a-a6c4-73d4f4133b95" containerID="90aebf031ec3f48a436b2122f40dd850b0eeee78920dc26741dd010da6c2f482" exitCode=0 Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.924385 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g4klr" event={"ID":"195700ca-d25a-470a-a6c4-73d4f4133b95","Type":"ContainerDied","Data":"90aebf031ec3f48a436b2122f40dd850b0eeee78920dc26741dd010da6c2f482"} Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.930915 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d612a58d-1418-42bb-b4d2-490ec2675699","Type":"ContainerStarted","Data":"c3703fc54abd8278a95bfe18de2525249a356e94701d27462c9b68cd3e4c323c"} Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.937581 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" event={"ID":"6a32cc27-948b-47d6-bd40-26360911278c","Type":"ContainerStarted","Data":"22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219"} Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.937768 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" podUID="6a32cc27-948b-47d6-bd40-26360911278c" containerName="dnsmasq-dns" containerID="cri-o://22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219" gracePeriod=10 Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.937938 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:35 crc kubenswrapper[4894]: I1208 15:04:35.974978 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" podStartSLOduration=3.974941368 podStartE2EDuration="3.974941368s" podCreationTimestamp="2025-12-08 15:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:35.962236668 +0000 UTC m=+1097.062242793" watchObservedRunningTime="2025-12-08 15:04:35.974941368 +0000 UTC m=+1097.074947493" Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.366268 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.443840 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.525980 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-8sqvg"] Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.768875 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6889cd597d-q9lvm"] Dec 08 15:04:36 crc kubenswrapper[4894]: W1208 15:04:36.782219 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bba85bc_cd5a_4b17_8dff_339cb0f745a7.slice/crio-780ffb924992ee6d559234abe082b4a97406b0071a3d409c81d514b7e28eef32 WatchSource:0}: Error finding container 780ffb924992ee6d559234abe082b4a97406b0071a3d409c81d514b7e28eef32: Status 404 returned error can't find the container with id 780ffb924992ee6d559234abe082b4a97406b0071a3d409c81d514b7e28eef32 Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.968585 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.973172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d612a58d-1418-42bb-b4d2-490ec2675699","Type":"ContainerStarted","Data":"8f9f024a19b768558dba7b19166baf0998f92d4c049e923d9bab5221b94a3c6f"} Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.975906 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6889cd597d-q9lvm" event={"ID":"4bba85bc-cd5a-4b17-8dff-339cb0f745a7","Type":"ContainerStarted","Data":"780ffb924992ee6d559234abe082b4a97406b0071a3d409c81d514b7e28eef32"} Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.978804 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" event={"ID":"8be96a69-9240-4484-96d0-8b5eb9e44ead","Type":"ContainerStarted","Data":"fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9"} Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.978846 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" event={"ID":"8be96a69-9240-4484-96d0-8b5eb9e44ead","Type":"ContainerStarted","Data":"8df4570b02e894d918705cb724fe3ce5a4e2ddf9aae23f36dcfa2395c80cf8d3"} Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.983164 4894 generic.go:334] "Generic (PLEG): container finished" podID="6a32cc27-948b-47d6-bd40-26360911278c" containerID="22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219" exitCode=0 Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.983233 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" event={"ID":"6a32cc27-948b-47d6-bd40-26360911278c","Type":"ContainerDied","Data":"22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219"} Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.983259 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" event={"ID":"6a32cc27-948b-47d6-bd40-26360911278c","Type":"ContainerDied","Data":"314305dea69d85b069af39bf7fda54cef35d5448b75c70f35bed00aeaaae632c"} Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.983281 4894 scope.go:117] "RemoveContainer" containerID="22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219" Dec 08 15:04:36 crc kubenswrapper[4894]: I1208 15:04:36.983294 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-hz8tj" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:36.995951 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5689da95-4134-41d1-9ab5-9c18f1995438","Type":"ContainerStarted","Data":"9633f9c2c01770279de3d05e8a175ad057c9698ed75b2b596ae63ed349607cd4"} Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.039794 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt7fr\" (UniqueName: \"kubernetes.io/projected/6a32cc27-948b-47d6-bd40-26360911278c-kube-api-access-vt7fr\") pod \"6a32cc27-948b-47d6-bd40-26360911278c\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.039923 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-nb\") pod \"6a32cc27-948b-47d6-bd40-26360911278c\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.040042 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-svc\") pod \"6a32cc27-948b-47d6-bd40-26360911278c\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.040120 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-sb\") pod \"6a32cc27-948b-47d6-bd40-26360911278c\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.040797 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-swift-storage-0\") pod \"6a32cc27-948b-47d6-bd40-26360911278c\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.040912 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-config\") pod \"6a32cc27-948b-47d6-bd40-26360911278c\" (UID: \"6a32cc27-948b-47d6-bd40-26360911278c\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.048127 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a32cc27-948b-47d6-bd40-26360911278c-kube-api-access-vt7fr" (OuterVolumeSpecName: "kube-api-access-vt7fr") pod "6a32cc27-948b-47d6-bd40-26360911278c" (UID: "6a32cc27-948b-47d6-bd40-26360911278c"). InnerVolumeSpecName "kube-api-access-vt7fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.103742 4894 scope.go:117] "RemoveContainer" containerID="d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.143369 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt7fr\" (UniqueName: \"kubernetes.io/projected/6a32cc27-948b-47d6-bd40-26360911278c-kube-api-access-vt7fr\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.160362 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a32cc27-948b-47d6-bd40-26360911278c" (UID: "6a32cc27-948b-47d6-bd40-26360911278c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.175245 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a32cc27-948b-47d6-bd40-26360911278c" (UID: "6a32cc27-948b-47d6-bd40-26360911278c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.190243 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a32cc27-948b-47d6-bd40-26360911278c" (UID: "6a32cc27-948b-47d6-bd40-26360911278c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.190748 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-config" (OuterVolumeSpecName: "config") pod "6a32cc27-948b-47d6-bd40-26360911278c" (UID: "6a32cc27-948b-47d6-bd40-26360911278c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.210833 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a32cc27-948b-47d6-bd40-26360911278c" (UID: "6a32cc27-948b-47d6-bd40-26360911278c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.270942 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.271020 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.271044 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.271060 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.271095 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a32cc27-948b-47d6-bd40-26360911278c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.409055 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-hz8tj"] Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.431758 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-hz8tj"] Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.657277 4894 scope.go:117] "RemoveContainer" containerID="22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219" Dec 08 15:04:37 crc kubenswrapper[4894]: E1208 15:04:37.685380 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219\": container with ID starting with 22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219 not found: ID does not exist" containerID="22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.685431 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219"} err="failed to get container status \"22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219\": rpc error: code = NotFound desc = could not find container \"22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219\": container with ID starting with 22081be63db65342e7c147d733355d4bd0fd104f32be054f1e48c96f04794219 not found: ID does not exist" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.685461 4894 scope.go:117] "RemoveContainer" containerID="d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c" Dec 08 15:04:37 crc kubenswrapper[4894]: E1208 15:04:37.686707 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c\": container with ID starting with d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c not found: ID does not exist" containerID="d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.687097 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c"} err="failed to get container status \"d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c\": rpc error: code = NotFound desc = could not find container \"d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c\": container with ID starting with d3207afa0ce63d44bf6551b50f2f2432d8745186e309a86c4676d0eb5545021c not found: ID does not exist" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.777239 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.914712 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-scripts\") pod \"195700ca-d25a-470a-a6c4-73d4f4133b95\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.915229 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-combined-ca-bundle\") pod \"195700ca-d25a-470a-a6c4-73d4f4133b95\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.915371 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-config-data\") pod \"195700ca-d25a-470a-a6c4-73d4f4133b95\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.915443 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znx6g\" (UniqueName: \"kubernetes.io/projected/195700ca-d25a-470a-a6c4-73d4f4133b95-kube-api-access-znx6g\") pod \"195700ca-d25a-470a-a6c4-73d4f4133b95\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.915476 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195700ca-d25a-470a-a6c4-73d4f4133b95-logs\") pod \"195700ca-d25a-470a-a6c4-73d4f4133b95\" (UID: \"195700ca-d25a-470a-a6c4-73d4f4133b95\") " Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.916230 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/195700ca-d25a-470a-a6c4-73d4f4133b95-logs" (OuterVolumeSpecName: "logs") pod "195700ca-d25a-470a-a6c4-73d4f4133b95" (UID: "195700ca-d25a-470a-a6c4-73d4f4133b95"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.922169 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-scripts" (OuterVolumeSpecName: "scripts") pod "195700ca-d25a-470a-a6c4-73d4f4133b95" (UID: "195700ca-d25a-470a-a6c4-73d4f4133b95"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.925056 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/195700ca-d25a-470a-a6c4-73d4f4133b95-kube-api-access-znx6g" (OuterVolumeSpecName: "kube-api-access-znx6g") pod "195700ca-d25a-470a-a6c4-73d4f4133b95" (UID: "195700ca-d25a-470a-a6c4-73d4f4133b95"). InnerVolumeSpecName "kube-api-access-znx6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:37 crc kubenswrapper[4894]: I1208 15:04:37.972971 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-config-data" (OuterVolumeSpecName: "config-data") pod "195700ca-d25a-470a-a6c4-73d4f4133b95" (UID: "195700ca-d25a-470a-a6c4-73d4f4133b95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.019411 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.019455 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.019467 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znx6g\" (UniqueName: \"kubernetes.io/projected/195700ca-d25a-470a-a6c4-73d4f4133b95-kube-api-access-znx6g\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.019477 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/195700ca-d25a-470a-a6c4-73d4f4133b95-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.052562 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "195700ca-d25a-470a-a6c4-73d4f4133b95" (UID: "195700ca-d25a-470a-a6c4-73d4f4133b95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.071684 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6f74bf4b48-xdrm8"] Dec 08 15:04:38 crc kubenswrapper[4894]: E1208 15:04:38.072392 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a32cc27-948b-47d6-bd40-26360911278c" containerName="init" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.072413 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a32cc27-948b-47d6-bd40-26360911278c" containerName="init" Dec 08 15:04:38 crc kubenswrapper[4894]: E1208 15:04:38.072448 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a32cc27-948b-47d6-bd40-26360911278c" containerName="dnsmasq-dns" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.072456 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a32cc27-948b-47d6-bd40-26360911278c" containerName="dnsmasq-dns" Dec 08 15:04:38 crc kubenswrapper[4894]: E1208 15:04:38.072477 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="195700ca-d25a-470a-a6c4-73d4f4133b95" containerName="placement-db-sync" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.072487 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="195700ca-d25a-470a-a6c4-73d4f4133b95" containerName="placement-db-sync" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.072761 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a32cc27-948b-47d6-bd40-26360911278c" containerName="dnsmasq-dns" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.072786 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="195700ca-d25a-470a-a6c4-73d4f4133b95" containerName="placement-db-sync" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.074382 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.078656 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.078990 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.095605 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g4klr" event={"ID":"195700ca-d25a-470a-a6c4-73d4f4133b95","Type":"ContainerDied","Data":"02412146543e671f6f92e070e4faa0ce4e09c3809a84e4eeef1659d66ef63d52"} Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.095674 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02412146543e671f6f92e070e4faa0ce4e09c3809a84e4eeef1659d66ef63d52" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.095841 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g4klr" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.099394 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f74bf4b48-xdrm8"] Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.100476 4894 generic.go:334] "Generic (PLEG): container finished" podID="77a9427b-9569-447d-aa60-e3be638e2d4c" containerID="d9a002ecf2a43b769b650f5eaf45af28c9c5bcf093c16c5beffca14bd9a79304" exitCode=0 Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.100547 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c642j" event={"ID":"77a9427b-9569-447d-aa60-e3be638e2d4c","Type":"ContainerDied","Data":"d9a002ecf2a43b769b650f5eaf45af28c9c5bcf093c16c5beffca14bd9a79304"} Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.113104 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6889cd597d-q9lvm" event={"ID":"4bba85bc-cd5a-4b17-8dff-339cb0f745a7","Type":"ContainerStarted","Data":"c43962dbee949786a1ae21324ab487aef33edfb2e0c1fa13add22deff95cda81"} Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.121079 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/195700ca-d25a-470a-a6c4-73d4f4133b95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.148323 4894 generic.go:334] "Generic (PLEG): container finished" podID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerID="fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9" exitCode=0 Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.148430 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" event={"ID":"8be96a69-9240-4484-96d0-8b5eb9e44ead","Type":"ContainerDied","Data":"fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9"} Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.230605 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-logs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.230688 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxz6p\" (UniqueName: \"kubernetes.io/projected/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-kube-api-access-mxz6p\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.230708 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-combined-ca-bundle\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.230793 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-internal-tls-certs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.230948 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-public-tls-certs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.230988 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-config-data\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.231020 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-scripts\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.332871 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-config-data\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.334130 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-scripts\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.334412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-logs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.334527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxz6p\" (UniqueName: \"kubernetes.io/projected/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-kube-api-access-mxz6p\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.334563 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-combined-ca-bundle\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.334859 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-internal-tls-certs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.335902 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-logs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.336468 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-public-tls-certs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.340200 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-combined-ca-bundle\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.340947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-public-tls-certs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.341366 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-scripts\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.343968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-internal-tls-certs\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.349103 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-config-data\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.368708 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxz6p\" (UniqueName: \"kubernetes.io/projected/cabb5ceb-58b0-47b5-8db0-cb717c6d8aad-kube-api-access-mxz6p\") pod \"placement-6f74bf4b48-xdrm8\" (UID: \"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad\") " pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.637152 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.890731 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5576bf896f-fxztt"] Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.893371 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.896392 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.901783 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 08 15:04:38 crc kubenswrapper[4894]: I1208 15:04:38.909555 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5576bf896f-fxztt"] Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.057883 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-combined-ca-bundle\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.057978 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-httpd-config\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.058009 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l657\" (UniqueName: \"kubernetes.io/projected/70163272-391d-4c33-be21-3e504f9a179f-kube-api-access-2l657\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.058044 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-config\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.058077 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-ovndb-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.058117 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-internal-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.058149 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-public-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.161015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-httpd-config\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.161307 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l657\" (UniqueName: \"kubernetes.io/projected/70163272-391d-4c33-be21-3e504f9a179f-kube-api-access-2l657\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.161435 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-config\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.161536 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-ovndb-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.161658 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-internal-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.161770 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-public-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.162003 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-combined-ca-bundle\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.170803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6889cd597d-q9lvm" event={"ID":"4bba85bc-cd5a-4b17-8dff-339cb0f745a7","Type":"ContainerStarted","Data":"02f2916c7384d69c38cfc90fa122b3d57d5b6a09dc89d3135906f2aa9067c004"} Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.172158 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.171088 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-combined-ca-bundle\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.172788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-config\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.173603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-internal-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.173748 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-ovndb-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.176671 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-public-tls-certs\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.178252 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" event={"ID":"8be96a69-9240-4484-96d0-8b5eb9e44ead","Type":"ContainerStarted","Data":"009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147"} Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.179377 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.181459 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/70163272-391d-4c33-be21-3e504f9a179f-httpd-config\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.186353 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l657\" (UniqueName: \"kubernetes.io/projected/70163272-391d-4c33-be21-3e504f9a179f-kube-api-access-2l657\") pod \"neutron-5576bf896f-fxztt\" (UID: \"70163272-391d-4c33-be21-3e504f9a179f\") " pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.188551 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-log" containerID="cri-o://9633f9c2c01770279de3d05e8a175ad057c9698ed75b2b596ae63ed349607cd4" gracePeriod=30 Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.188698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5689da95-4134-41d1-9ab5-9c18f1995438","Type":"ContainerStarted","Data":"c594e0c5dcc16df2335f1449b117bfd30533198a74973942f452ee355b0662f7"} Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.188799 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-httpd" containerID="cri-o://c594e0c5dcc16df2335f1449b117bfd30533198a74973942f452ee355b0662f7" gracePeriod=30 Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.209731 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6889cd597d-q9lvm" podStartSLOduration=4.209699672 podStartE2EDuration="4.209699672s" podCreationTimestamp="2025-12-08 15:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:39.201283807 +0000 UTC m=+1100.301289932" watchObservedRunningTime="2025-12-08 15:04:39.209699672 +0000 UTC m=+1100.309705787" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.218605 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-log" containerID="cri-o://8f9f024a19b768558dba7b19166baf0998f92d4c049e923d9bab5221b94a3c6f" gracePeriod=30 Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.219160 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-httpd" containerID="cri-o://365413096f6bbe91d80231d8165e17fcb7eeb226916f4c7663aeb99d0191bb70" gracePeriod=30 Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.222363 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a32cc27-948b-47d6-bd40-26360911278c" path="/var/lib/kubelet/pods/6a32cc27-948b-47d6-bd40-26360911278c/volumes" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.223333 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d612a58d-1418-42bb-b4d2-490ec2675699","Type":"ContainerStarted","Data":"365413096f6bbe91d80231d8165e17fcb7eeb226916f4c7663aeb99d0191bb70"} Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.236348 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.245531 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.245504827 podStartE2EDuration="6.245504827s" podCreationTimestamp="2025-12-08 15:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:39.237684142 +0000 UTC m=+1100.337690257" watchObservedRunningTime="2025-12-08 15:04:39.245504827 +0000 UTC m=+1100.345510942" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.289531 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" podStartSLOduration=4.289493521 podStartE2EDuration="4.289493521s" podCreationTimestamp="2025-12-08 15:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:39.275163501 +0000 UTC m=+1100.375169616" watchObservedRunningTime="2025-12-08 15:04:39.289493521 +0000 UTC m=+1100.389499636" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.309322 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f74bf4b48-xdrm8"] Dec 08 15:04:39 crc kubenswrapper[4894]: W1208 15:04:39.348709 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcabb5ceb_58b0_47b5_8db0_cb717c6d8aad.slice/crio-8dc2ddd2145a8612cd3344ccaebb19d751e593a4449d9e488d758c8cdb1f9f8f WatchSource:0}: Error finding container 8dc2ddd2145a8612cd3344ccaebb19d751e593a4449d9e488d758c8cdb1f9f8f: Status 404 returned error can't find the container with id 8dc2ddd2145a8612cd3344ccaebb19d751e593a4449d9e488d758c8cdb1f9f8f Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.748626 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.778636 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.778613833 podStartE2EDuration="7.778613833s" podCreationTimestamp="2025-12-08 15:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:39.329729826 +0000 UTC m=+1100.429735941" watchObservedRunningTime="2025-12-08 15:04:39.778613833 +0000 UTC m=+1100.878619948" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.888878 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-config-data\") pod \"77a9427b-9569-447d-aa60-e3be638e2d4c\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.889040 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-credential-keys\") pod \"77a9427b-9569-447d-aa60-e3be638e2d4c\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.889175 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-scripts\") pod \"77a9427b-9569-447d-aa60-e3be638e2d4c\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.889210 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-fernet-keys\") pod \"77a9427b-9569-447d-aa60-e3be638e2d4c\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.889286 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-combined-ca-bundle\") pod \"77a9427b-9569-447d-aa60-e3be638e2d4c\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.889314 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76qcv\" (UniqueName: \"kubernetes.io/projected/77a9427b-9569-447d-aa60-e3be638e2d4c-kube-api-access-76qcv\") pod \"77a9427b-9569-447d-aa60-e3be638e2d4c\" (UID: \"77a9427b-9569-447d-aa60-e3be638e2d4c\") " Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.898254 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a9427b-9569-447d-aa60-e3be638e2d4c-kube-api-access-76qcv" (OuterVolumeSpecName: "kube-api-access-76qcv") pod "77a9427b-9569-447d-aa60-e3be638e2d4c" (UID: "77a9427b-9569-447d-aa60-e3be638e2d4c"). InnerVolumeSpecName "kube-api-access-76qcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.914161 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "77a9427b-9569-447d-aa60-e3be638e2d4c" (UID: "77a9427b-9569-447d-aa60-e3be638e2d4c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.916016 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "77a9427b-9569-447d-aa60-e3be638e2d4c" (UID: "77a9427b-9569-447d-aa60-e3be638e2d4c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.916175 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-scripts" (OuterVolumeSpecName: "scripts") pod "77a9427b-9569-447d-aa60-e3be638e2d4c" (UID: "77a9427b-9569-447d-aa60-e3be638e2d4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.972447 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-config-data" (OuterVolumeSpecName: "config-data") pod "77a9427b-9569-447d-aa60-e3be638e2d4c" (UID: "77a9427b-9569-447d-aa60-e3be638e2d4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.994009 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77a9427b-9569-447d-aa60-e3be638e2d4c" (UID: "77a9427b-9569-447d-aa60-e3be638e2d4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.996870 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.996922 4894 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.996938 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.996948 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.996959 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77a9427b-9569-447d-aa60-e3be638e2d4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:39 crc kubenswrapper[4894]: I1208 15:04:39.996975 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76qcv\" (UniqueName: \"kubernetes.io/projected/77a9427b-9569-447d-aa60-e3be638e2d4c-kube-api-access-76qcv\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: W1208 15:04:40.097936 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70163272_391d_4c33_be21_3e504f9a179f.slice/crio-e47c909373363718d23cc6ba67a8d1975f7fd92baa3f035364470d54a6379fc7 WatchSource:0}: Error finding container e47c909373363718d23cc6ba67a8d1975f7fd92baa3f035364470d54a6379fc7: Status 404 returned error can't find the container with id e47c909373363718d23cc6ba67a8d1975f7fd92baa3f035364470d54a6379fc7 Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.098103 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5576bf896f-fxztt"] Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.172760 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.172837 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.267640 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7cb79b8b98-42mc8"] Dec 08 15:04:40 crc kubenswrapper[4894]: E1208 15:04:40.268827 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a9427b-9569-447d-aa60-e3be638e2d4c" containerName="keystone-bootstrap" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.268853 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a9427b-9569-447d-aa60-e3be638e2d4c" containerName="keystone-bootstrap" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.269140 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a9427b-9569-447d-aa60-e3be638e2d4c" containerName="keystone-bootstrap" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.277544 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.283491 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.283747 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.300013 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.305553 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.315107 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7cb79b8b98-42mc8"] Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.326843 4894 generic.go:334] "Generic (PLEG): container finished" podID="5689da95-4134-41d1-9ab5-9c18f1995438" containerID="c594e0c5dcc16df2335f1449b117bfd30533198a74973942f452ee355b0662f7" exitCode=0 Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.326898 4894 generic.go:334] "Generic (PLEG): container finished" podID="5689da95-4134-41d1-9ab5-9c18f1995438" containerID="9633f9c2c01770279de3d05e8a175ad057c9698ed75b2b596ae63ed349607cd4" exitCode=143 Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.327020 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5689da95-4134-41d1-9ab5-9c18f1995438","Type":"ContainerDied","Data":"c594e0c5dcc16df2335f1449b117bfd30533198a74973942f452ee355b0662f7"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.327064 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5689da95-4134-41d1-9ab5-9c18f1995438","Type":"ContainerDied","Data":"9633f9c2c01770279de3d05e8a175ad057c9698ed75b2b596ae63ed349607cd4"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.327075 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5689da95-4134-41d1-9ab5-9c18f1995438","Type":"ContainerDied","Data":"e88539449f678d9e373346cb6c973a0d822cbdeefd5ed38f38298a354ea62e75"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.327086 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e88539449f678d9e373346cb6c973a0d822cbdeefd5ed38f38298a354ea62e75" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.342325 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-c642j" event={"ID":"77a9427b-9569-447d-aa60-e3be638e2d4c","Type":"ContainerDied","Data":"e718982841f24ff257045e10dc0fd07d57cc3c68d60eb78e5a394d703d3fa7c3"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.342380 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e718982841f24ff257045e10dc0fd07d57cc3c68d60eb78e5a394d703d3fa7c3" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.342452 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-c642j" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.360547 4894 generic.go:334] "Generic (PLEG): container finished" podID="d612a58d-1418-42bb-b4d2-490ec2675699" containerID="365413096f6bbe91d80231d8165e17fcb7eeb226916f4c7663aeb99d0191bb70" exitCode=0 Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.360605 4894 generic.go:334] "Generic (PLEG): container finished" podID="d612a58d-1418-42bb-b4d2-490ec2675699" containerID="8f9f024a19b768558dba7b19166baf0998f92d4c049e923d9bab5221b94a3c6f" exitCode=143 Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.360667 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d612a58d-1418-42bb-b4d2-490ec2675699","Type":"ContainerDied","Data":"365413096f6bbe91d80231d8165e17fcb7eeb226916f4c7663aeb99d0191bb70"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.360707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d612a58d-1418-42bb-b4d2-490ec2675699","Type":"ContainerDied","Data":"8f9f024a19b768558dba7b19166baf0998f92d4c049e923d9bab5221b94a3c6f"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.361783 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.363142 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5576bf896f-fxztt" event={"ID":"70163272-391d-4c33-be21-3e504f9a179f","Type":"ContainerStarted","Data":"e47c909373363718d23cc6ba67a8d1975f7fd92baa3f035364470d54a6379fc7"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.371406 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f74bf4b48-xdrm8" event={"ID":"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad","Type":"ContainerStarted","Data":"c7b898fc1add461b8ad501fa4409b0d50ff166e20af4a0fc092c6242d25667a8"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.371442 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.371455 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f74bf4b48-xdrm8" event={"ID":"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad","Type":"ContainerStarted","Data":"6ee2109422405b0613a3c5d11137d97929e996cea0564a7c15e35db892bb7a23"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.371466 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f74bf4b48-xdrm8" event={"ID":"cabb5ceb-58b0-47b5-8db0-cb717c6d8aad","Type":"ContainerStarted","Data":"8dc2ddd2145a8612cd3344ccaebb19d751e593a4449d9e488d758c8cdb1f9f8f"} Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.371480 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.429919 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"5689da95-4134-41d1-9ab5-9c18f1995438\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430127 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-httpd-run\") pod \"5689da95-4134-41d1-9ab5-9c18f1995438\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430183 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-scripts\") pod \"5689da95-4134-41d1-9ab5-9c18f1995438\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430206 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs2bv\" (UniqueName: \"kubernetes.io/projected/5689da95-4134-41d1-9ab5-9c18f1995438-kube-api-access-zs2bv\") pod \"5689da95-4134-41d1-9ab5-9c18f1995438\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430416 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-combined-ca-bundle\") pod \"5689da95-4134-41d1-9ab5-9c18f1995438\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430502 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-logs\") pod \"5689da95-4134-41d1-9ab5-9c18f1995438\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430537 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-config-data\") pod \"5689da95-4134-41d1-9ab5-9c18f1995438\" (UID: \"5689da95-4134-41d1-9ab5-9c18f1995438\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430834 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htmnb\" (UniqueName: \"kubernetes.io/projected/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-kube-api-access-htmnb\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430884 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-combined-ca-bundle\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430926 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-internal-tls-certs\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.430945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-credential-keys\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.431017 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-public-tls-certs\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.431106 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-scripts\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.431143 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-config-data\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.431427 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-fernet-keys\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.435489 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-logs" (OuterVolumeSpecName: "logs") pod "5689da95-4134-41d1-9ab5-9c18f1995438" (UID: "5689da95-4134-41d1-9ab5-9c18f1995438"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.436238 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5689da95-4134-41d1-9ab5-9c18f1995438" (UID: "5689da95-4134-41d1-9ab5-9c18f1995438"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.440013 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-scripts" (OuterVolumeSpecName: "scripts") pod "5689da95-4134-41d1-9ab5-9c18f1995438" (UID: "5689da95-4134-41d1-9ab5-9c18f1995438"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.440033 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5689da95-4134-41d1-9ab5-9c18f1995438-kube-api-access-zs2bv" (OuterVolumeSpecName: "kube-api-access-zs2bv") pod "5689da95-4134-41d1-9ab5-9c18f1995438" (UID: "5689da95-4134-41d1-9ab5-9c18f1995438"). InnerVolumeSpecName "kube-api-access-zs2bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.448860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "5689da95-4134-41d1-9ab5-9c18f1995438" (UID: "5689da95-4134-41d1-9ab5-9c18f1995438"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.455726 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6f74bf4b48-xdrm8" podStartSLOduration=2.455703375 podStartE2EDuration="2.455703375s" podCreationTimestamp="2025-12-08 15:04:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:40.450113649 +0000 UTC m=+1101.550119774" watchObservedRunningTime="2025-12-08 15:04:40.455703375 +0000 UTC m=+1101.555709490" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.491450 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5689da95-4134-41d1-9ab5-9c18f1995438" (UID: "5689da95-4134-41d1-9ab5-9c18f1995438"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.519641 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534404 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-fernet-keys\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534494 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htmnb\" (UniqueName: \"kubernetes.io/projected/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-kube-api-access-htmnb\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534522 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-combined-ca-bundle\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534555 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-internal-tls-certs\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534576 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-credential-keys\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534610 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-public-tls-certs\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534934 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-scripts\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.534966 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-config-data\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.535050 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.535066 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.535075 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.535086 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs2bv\" (UniqueName: \"kubernetes.io/projected/5689da95-4134-41d1-9ab5-9c18f1995438-kube-api-access-zs2bv\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.535098 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.535107 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5689da95-4134-41d1-9ab5-9c18f1995438-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.543245 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-scripts\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.544692 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-fernet-keys\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.549054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-combined-ca-bundle\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.549805 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-config-data" (OuterVolumeSpecName: "config-data") pod "5689da95-4134-41d1-9ab5-9c18f1995438" (UID: "5689da95-4134-41d1-9ab5-9c18f1995438"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.551756 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-credential-keys\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.553321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-public-tls-certs\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.558347 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-config-data\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.560234 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-internal-tls-certs\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.564665 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htmnb\" (UniqueName: \"kubernetes.io/projected/57827c7d-7ca8-4eb4-80e9-c6690fca1dac-kube-api-access-htmnb\") pod \"keystone-7cb79b8b98-42mc8\" (UID: \"57827c7d-7ca8-4eb4-80e9-c6690fca1dac\") " pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.624581 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.639613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcr6d\" (UniqueName: \"kubernetes.io/projected/d612a58d-1418-42bb-b4d2-490ec2675699-kube-api-access-hcr6d\") pod \"d612a58d-1418-42bb-b4d2-490ec2675699\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.639709 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d612a58d-1418-42bb-b4d2-490ec2675699\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.639856 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-logs\") pod \"d612a58d-1418-42bb-b4d2-490ec2675699\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.639902 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-scripts\") pod \"d612a58d-1418-42bb-b4d2-490ec2675699\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.639978 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-combined-ca-bundle\") pod \"d612a58d-1418-42bb-b4d2-490ec2675699\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.640011 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-config-data\") pod \"d612a58d-1418-42bb-b4d2-490ec2675699\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.640124 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-httpd-run\") pod \"d612a58d-1418-42bb-b4d2-490ec2675699\" (UID: \"d612a58d-1418-42bb-b4d2-490ec2675699\") " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.642624 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-logs" (OuterVolumeSpecName: "logs") pod "d612a58d-1418-42bb-b4d2-490ec2675699" (UID: "d612a58d-1418-42bb-b4d2-490ec2675699"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.650595 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "d612a58d-1418-42bb-b4d2-490ec2675699" (UID: "d612a58d-1418-42bb-b4d2-490ec2675699"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.651168 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d612a58d-1418-42bb-b4d2-490ec2675699-kube-api-access-hcr6d" (OuterVolumeSpecName: "kube-api-access-hcr6d") pod "d612a58d-1418-42bb-b4d2-490ec2675699" (UID: "d612a58d-1418-42bb-b4d2-490ec2675699"). InnerVolumeSpecName "kube-api-access-hcr6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.652926 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5689da95-4134-41d1-9ab5-9c18f1995438-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.652957 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.661163 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.661494 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d612a58d-1418-42bb-b4d2-490ec2675699" (UID: "d612a58d-1418-42bb-b4d2-490ec2675699"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.661826 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-scripts" (OuterVolumeSpecName: "scripts") pod "d612a58d-1418-42bb-b4d2-490ec2675699" (UID: "d612a58d-1418-42bb-b4d2-490ec2675699"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.726176 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d612a58d-1418-42bb-b4d2-490ec2675699" (UID: "d612a58d-1418-42bb-b4d2-490ec2675699"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.754914 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcr6d\" (UniqueName: \"kubernetes.io/projected/d612a58d-1418-42bb-b4d2-490ec2675699-kube-api-access-hcr6d\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.754982 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.754995 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.755004 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.755016 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.755025 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d612a58d-1418-42bb-b4d2-490ec2675699-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.782008 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.795019 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-config-data" (OuterVolumeSpecName: "config-data") pod "d612a58d-1418-42bb-b4d2-490ec2675699" (UID: "d612a58d-1418-42bb-b4d2-490ec2675699"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.858605 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d612a58d-1418-42bb-b4d2-490ec2675699-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:40 crc kubenswrapper[4894]: I1208 15:04:40.858953 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.291125 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7cb79b8b98-42mc8"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.384851 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d612a58d-1418-42bb-b4d2-490ec2675699","Type":"ContainerDied","Data":"c3703fc54abd8278a95bfe18de2525249a356e94701d27462c9b68cd3e4c323c"} Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.385126 4894 scope.go:117] "RemoveContainer" containerID="365413096f6bbe91d80231d8165e17fcb7eeb226916f4c7663aeb99d0191bb70" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.385018 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.391512 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5576bf896f-fxztt" event={"ID":"70163272-391d-4c33-be21-3e504f9a179f","Type":"ContainerStarted","Data":"bb10261756c76bad2e47f772065937bf852f61dc5d21d4c214b9ed2f917c86aa"} Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.391575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5576bf896f-fxztt" event={"ID":"70163272-391d-4c33-be21-3e504f9a179f","Type":"ContainerStarted","Data":"ecd0ec5a140fea79d9ba76567dead380195da47622019e6b11865ec511b8ad36"} Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.391783 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.391888 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.435229 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5576bf896f-fxztt" podStartSLOduration=3.435193153 podStartE2EDuration="3.435193153s" podCreationTimestamp="2025-12-08 15:04:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:41.413114092 +0000 UTC m=+1102.513120227" watchObservedRunningTime="2025-12-08 15:04:41.435193153 +0000 UTC m=+1102.535199268" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.461032 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.493063 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.508947 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.591091 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: E1208 15:04:41.591696 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-log" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.591711 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-log" Dec 08 15:04:41 crc kubenswrapper[4894]: E1208 15:04:41.591737 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-log" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.591743 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-log" Dec 08 15:04:41 crc kubenswrapper[4894]: E1208 15:04:41.591760 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-httpd" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.591767 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-httpd" Dec 08 15:04:41 crc kubenswrapper[4894]: E1208 15:04:41.591775 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-httpd" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.591781 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-httpd" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.592011 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-log" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.592033 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-log" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.592049 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" containerName="glance-httpd" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.592059 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" containerName="glance-httpd" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.593429 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.600512 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.600738 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lg9b9" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.600745 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.601115 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.630922 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.663028 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.681631 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.683325 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.691413 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.691649 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705608 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-scripts\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705679 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-config-data\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705751 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh228\" (UniqueName: \"kubernetes.io/projected/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-kube-api-access-gh228\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705787 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705834 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705875 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705902 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.705965 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-logs\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.711478 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808079 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808157 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-scripts\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808198 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808227 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-logs\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808288 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-config-data\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808335 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808358 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-logs\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808380 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808410 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmnt5\" (UniqueName: \"kubernetes.io/projected/be1f50cd-10f7-441e-a140-63713dcafc2d-kube-api-access-wmnt5\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808441 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808477 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-scripts\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808553 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-config-data\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808628 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh228\" (UniqueName: \"kubernetes.io/projected/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-kube-api-access-gh228\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808665 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.808694 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.809374 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.811042 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.828262 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-logs\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.833637 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-scripts\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.834564 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.843075 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-config-data\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.851232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh228\" (UniqueName: \"kubernetes.io/projected/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-kube-api-access-gh228\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.864740 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.870464 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.917969 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-scripts\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.918049 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-logs\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.918089 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-config-data\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.918119 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.918142 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.918162 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmnt5\" (UniqueName: \"kubernetes.io/projected/be1f50cd-10f7-441e-a140-63713dcafc2d-kube-api-access-wmnt5\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.918187 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.918212 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.937591 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.939567 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.947470 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.947975 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.948239 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-logs\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.954390 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-scripts\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.960550 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-config-data\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.975805 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:41 crc kubenswrapper[4894]: I1208 15:04:41.989408 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmnt5\" (UniqueName: \"kubernetes.io/projected/be1f50cd-10f7-441e-a140-63713dcafc2d-kube-api-access-wmnt5\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:42 crc kubenswrapper[4894]: I1208 15:04:42.018414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " pod="openstack/glance-default-external-api-0" Dec 08 15:04:42 crc kubenswrapper[4894]: I1208 15:04:42.316259 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:04:43 crc kubenswrapper[4894]: I1208 15:04:43.216304 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5689da95-4134-41d1-9ab5-9c18f1995438" path="/var/lib/kubelet/pods/5689da95-4134-41d1-9ab5-9c18f1995438/volumes" Dec 08 15:04:43 crc kubenswrapper[4894]: I1208 15:04:43.221680 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d612a58d-1418-42bb-b4d2-490ec2675699" path="/var/lib/kubelet/pods/d612a58d-1418-42bb-b4d2-490ec2675699/volumes" Dec 08 15:04:45 crc kubenswrapper[4894]: I1208 15:04:45.753011 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:04:45 crc kubenswrapper[4894]: I1208 15:04:45.840153 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-2mjnx"] Dec 08 15:04:45 crc kubenswrapper[4894]: I1208 15:04:45.840389 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerName="dnsmasq-dns" containerID="cri-o://6e87f0ad99b65d8329594708817c1d0806b34a4d0a524d037510213144f1924c" gracePeriod=10 Dec 08 15:04:45 crc kubenswrapper[4894]: I1208 15:04:45.899651 4894 scope.go:117] "RemoveContainer" containerID="8f9f024a19b768558dba7b19166baf0998f92d4c049e923d9bab5221b94a3c6f" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.585250 4894 generic.go:334] "Generic (PLEG): container finished" podID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerID="6e87f0ad99b65d8329594708817c1d0806b34a4d0a524d037510213144f1924c" exitCode=0 Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.585710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" event={"ID":"97e76d4c-be0b-4ed9-8e51-c5bed27d5514","Type":"ContainerDied","Data":"6e87f0ad99b65d8329594708817c1d0806b34a4d0a524d037510213144f1924c"} Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.593004 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7cb79b8b98-42mc8" event={"ID":"57827c7d-7ca8-4eb4-80e9-c6690fca1dac","Type":"ContainerStarted","Data":"93112573d0e6aa9639561cdb1b29233361026fb95de282f5515b22e8595f5949"} Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.593719 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.642873 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7cb79b8b98-42mc8" podStartSLOduration=6.64283281 podStartE2EDuration="6.64283281s" podCreationTimestamp="2025-12-08 15:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:46.625073735 +0000 UTC m=+1107.725079850" watchObservedRunningTime="2025-12-08 15:04:46.64283281 +0000 UTC m=+1107.742838935" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.678110 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.845666 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57vq2\" (UniqueName: \"kubernetes.io/projected/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-kube-api-access-57vq2\") pod \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.845750 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-config\") pod \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.845892 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-svc\") pod \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.846009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-swift-storage-0\") pod \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.846065 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-nb\") pod \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.846117 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-sb\") pod \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\" (UID: \"97e76d4c-be0b-4ed9-8e51-c5bed27d5514\") " Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.865224 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-kube-api-access-57vq2" (OuterVolumeSpecName: "kube-api-access-57vq2") pod "97e76d4c-be0b-4ed9-8e51-c5bed27d5514" (UID: "97e76d4c-be0b-4ed9-8e51-c5bed27d5514"). InnerVolumeSpecName "kube-api-access-57vq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.929663 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "97e76d4c-be0b-4ed9-8e51-c5bed27d5514" (UID: "97e76d4c-be0b-4ed9-8e51-c5bed27d5514"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.948878 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57vq2\" (UniqueName: \"kubernetes.io/projected/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-kube-api-access-57vq2\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.948907 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.967381 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "97e76d4c-be0b-4ed9-8e51-c5bed27d5514" (UID: "97e76d4c-be0b-4ed9-8e51-c5bed27d5514"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.968299 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-config" (OuterVolumeSpecName: "config") pod "97e76d4c-be0b-4ed9-8e51-c5bed27d5514" (UID: "97e76d4c-be0b-4ed9-8e51-c5bed27d5514"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:46 crc kubenswrapper[4894]: I1208 15:04:46.968417 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "97e76d4c-be0b-4ed9-8e51-c5bed27d5514" (UID: "97e76d4c-be0b-4ed9-8e51-c5bed27d5514"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.002779 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "97e76d4c-be0b-4ed9-8e51-c5bed27d5514" (UID: "97e76d4c-be0b-4ed9-8e51-c5bed27d5514"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.054895 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.054939 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.054954 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.054965 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/97e76d4c-be0b-4ed9-8e51-c5bed27d5514-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.075908 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.608946 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7nxjg" event={"ID":"06177501-5baa-49e0-94bd-157c6f079c05","Type":"ContainerStarted","Data":"3daa8390366acee3732b5671a4160901220b017445b783e7d47984580a880039"} Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.612691 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" event={"ID":"97e76d4c-be0b-4ed9-8e51-c5bed27d5514","Type":"ContainerDied","Data":"f800e34eeb710ed4ea1b663212a1f495a517d4073fecce9547f742f6f70a162a"} Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.612745 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.612759 4894 scope.go:117] "RemoveContainer" containerID="6e87f0ad99b65d8329594708817c1d0806b34a4d0a524d037510213144f1924c" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.616692 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7cb79b8b98-42mc8" event={"ID":"57827c7d-7ca8-4eb4-80e9-c6690fca1dac","Type":"ContainerStarted","Data":"cd52f716a2c0e1a896a88a7f41fe796d6e93b5681c377eacfaae8512168da59f"} Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.620800 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerStarted","Data":"d5a83a8baf3ba09d4e59a5447a2b742b13cb1f8293257083f73de6104b9d931b"} Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.623350 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27","Type":"ContainerStarted","Data":"b537dfa27b314283be1f07456800575b37d24bba219f33649bc694eecdc42f4b"} Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.638725 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7b8gb" event={"ID":"85f57102-ceaf-484a-a0b3-c1ecf324339d","Type":"ContainerStarted","Data":"ef31a8436b105850224ef08c14f32232599779e1eabc7a8b36d5df4cf16c35a2"} Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.651847 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7nxjg" podStartSLOduration=3.348957489 podStartE2EDuration="47.651804341s" podCreationTimestamp="2025-12-08 15:04:00 +0000 UTC" firstStartedPulling="2025-12-08 15:04:01.977787045 +0000 UTC m=+1063.077793160" lastFinishedPulling="2025-12-08 15:04:46.280633897 +0000 UTC m=+1107.380640012" observedRunningTime="2025-12-08 15:04:47.633385233 +0000 UTC m=+1108.733391348" watchObservedRunningTime="2025-12-08 15:04:47.651804341 +0000 UTC m=+1108.751810456" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.691153 4894 scope.go:117] "RemoveContainer" containerID="043a6234a41046010426ee7943ab02197e6cf810b16af6279d11596187deee29" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.692909 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-2mjnx"] Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.700919 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-2mjnx"] Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.706752 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-7b8gb" podStartSLOduration=3.911777251 podStartE2EDuration="47.70672856s" podCreationTimestamp="2025-12-08 15:04:00 +0000 UTC" firstStartedPulling="2025-12-08 15:04:02.491411318 +0000 UTC m=+1063.591417433" lastFinishedPulling="2025-12-08 15:04:46.286362627 +0000 UTC m=+1107.386368742" observedRunningTime="2025-12-08 15:04:47.666985256 +0000 UTC m=+1108.766991361" watchObservedRunningTime="2025-12-08 15:04:47.70672856 +0000 UTC m=+1108.806734675" Dec 08 15:04:47 crc kubenswrapper[4894]: I1208 15:04:47.834960 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:04:48 crc kubenswrapper[4894]: I1208 15:04:48.663140 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27","Type":"ContainerStarted","Data":"02a88beb76b7aedf12add8d09eea22fcbe2316c0e26a00b526674848370c6231"} Dec 08 15:04:48 crc kubenswrapper[4894]: I1208 15:04:48.666969 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be1f50cd-10f7-441e-a140-63713dcafc2d","Type":"ContainerStarted","Data":"61aee1f2c5b5dd7f73ee5ee3e12e85e30912b3e9ed28d3adba8981ebab0bcde6"} Dec 08 15:04:49 crc kubenswrapper[4894]: I1208 15:04:49.215164 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" path="/var/lib/kubelet/pods/97e76d4c-be0b-4ed9-8e51-c5bed27d5514/volumes" Dec 08 15:04:49 crc kubenswrapper[4894]: I1208 15:04:49.685596 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be1f50cd-10f7-441e-a140-63713dcafc2d","Type":"ContainerStarted","Data":"45b362505a0bb89c90c563f29cab56658bb7e4405698d6e4eb3a96702676892a"} Dec 08 15:04:49 crc kubenswrapper[4894]: I1208 15:04:49.685955 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be1f50cd-10f7-441e-a140-63713dcafc2d","Type":"ContainerStarted","Data":"ec3e156e176ec3b29b8d573109a6e2c1c7478ff41628a3698b819bac263d21f0"} Dec 08 15:04:49 crc kubenswrapper[4894]: I1208 15:04:49.689888 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27","Type":"ContainerStarted","Data":"eab57615ea0b8264dc40b7e61bd46025b147db181ff921e453864db32dae66d8"} Dec 08 15:04:49 crc kubenswrapper[4894]: I1208 15:04:49.746481 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.746461663 podStartE2EDuration="8.746461663s" podCreationTimestamp="2025-12-08 15:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:49.719173388 +0000 UTC m=+1110.819179503" watchObservedRunningTime="2025-12-08 15:04:49.746461663 +0000 UTC m=+1110.846467778" Dec 08 15:04:49 crc kubenswrapper[4894]: I1208 15:04:49.781093 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.781067606 podStartE2EDuration="8.781067606s" podCreationTimestamp="2025-12-08 15:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:49.751217872 +0000 UTC m=+1110.851223997" watchObservedRunningTime="2025-12-08 15:04:49.781067606 +0000 UTC m=+1110.881073731" Dec 08 15:04:50 crc kubenswrapper[4894]: I1208 15:04:50.171977 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7fb57bf44-rk2jn" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 08 15:04:50 crc kubenswrapper[4894]: I1208 15:04:50.302218 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-797dcfbfb8-294k4" podUID="c3009b99-5e3b-4f39-b62f-0e04aeb371f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 08 15:04:51 crc kubenswrapper[4894]: I1208 15:04:51.394123 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-2mjnx" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Dec 08 15:04:51 crc kubenswrapper[4894]: I1208 15:04:51.714994 4894 generic.go:334] "Generic (PLEG): container finished" podID="85f57102-ceaf-484a-a0b3-c1ecf324339d" containerID="ef31a8436b105850224ef08c14f32232599779e1eabc7a8b36d5df4cf16c35a2" exitCode=0 Dec 08 15:04:51 crc kubenswrapper[4894]: I1208 15:04:51.715111 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7b8gb" event={"ID":"85f57102-ceaf-484a-a0b3-c1ecf324339d","Type":"ContainerDied","Data":"ef31a8436b105850224ef08c14f32232599779e1eabc7a8b36d5df4cf16c35a2"} Dec 08 15:04:51 crc kubenswrapper[4894]: I1208 15:04:51.950149 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:51 crc kubenswrapper[4894]: I1208 15:04:51.950196 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:51 crc kubenswrapper[4894]: I1208 15:04:51.997428 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.009366 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.316751 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.319402 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.368238 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.368353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.733594 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.733651 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.733666 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 15:04:52 crc kubenswrapper[4894]: I1208 15:04:52.733683 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 15:04:54 crc kubenswrapper[4894]: I1208 15:04:54.756991 4894 generic.go:334] "Generic (PLEG): container finished" podID="06177501-5baa-49e0-94bd-157c6f079c05" containerID="3daa8390366acee3732b5671a4160901220b017445b783e7d47984580a880039" exitCode=0 Dec 08 15:04:54 crc kubenswrapper[4894]: I1208 15:04:54.757035 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7nxjg" event={"ID":"06177501-5baa-49e0-94bd-157c6f079c05","Type":"ContainerDied","Data":"3daa8390366acee3732b5671a4160901220b017445b783e7d47984580a880039"} Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.085448 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.086022 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.456016 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.502772 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5z2l\" (UniqueName: \"kubernetes.io/projected/85f57102-ceaf-484a-a0b3-c1ecf324339d-kube-api-access-g5z2l\") pod \"85f57102-ceaf-484a-a0b3-c1ecf324339d\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.502983 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-combined-ca-bundle\") pod \"85f57102-ceaf-484a-a0b3-c1ecf324339d\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.503159 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-db-sync-config-data\") pod \"85f57102-ceaf-484a-a0b3-c1ecf324339d\" (UID: \"85f57102-ceaf-484a-a0b3-c1ecf324339d\") " Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.509704 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "85f57102-ceaf-484a-a0b3-c1ecf324339d" (UID: "85f57102-ceaf-484a-a0b3-c1ecf324339d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.509760 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85f57102-ceaf-484a-a0b3-c1ecf324339d-kube-api-access-g5z2l" (OuterVolumeSpecName: "kube-api-access-g5z2l") pod "85f57102-ceaf-484a-a0b3-c1ecf324339d" (UID: "85f57102-ceaf-484a-a0b3-c1ecf324339d"). InnerVolumeSpecName "kube-api-access-g5z2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.603446 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85f57102-ceaf-484a-a0b3-c1ecf324339d" (UID: "85f57102-ceaf-484a-a0b3-c1ecf324339d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.605754 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.605793 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5z2l\" (UniqueName: \"kubernetes.io/projected/85f57102-ceaf-484a-a0b3-c1ecf324339d-kube-api-access-g5z2l\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.606037 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f57102-ceaf-484a-a0b3-c1ecf324339d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.774138 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7b8gb" Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.774191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7b8gb" event={"ID":"85f57102-ceaf-484a-a0b3-c1ecf324339d","Type":"ContainerDied","Data":"c3ab14a8234828cbf2885308e455f4dd432abb6a0b75092c3d53f3b35b7b5084"} Dec 08 15:04:55 crc kubenswrapper[4894]: I1208 15:04:55.774234 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3ab14a8234828cbf2885308e455f4dd432abb6a0b75092c3d53f3b35b7b5084" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.316679 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.425306 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06177501-5baa-49e0-94bd-157c6f079c05-etc-machine-id\") pod \"06177501-5baa-49e0-94bd-157c6f079c05\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.425420 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06177501-5baa-49e0-94bd-157c6f079c05-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "06177501-5baa-49e0-94bd-157c6f079c05" (UID: "06177501-5baa-49e0-94bd-157c6f079c05"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.425488 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-combined-ca-bundle\") pod \"06177501-5baa-49e0-94bd-157c6f079c05\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.425517 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-db-sync-config-data\") pod \"06177501-5baa-49e0-94bd-157c6f079c05\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.425541 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-config-data\") pod \"06177501-5baa-49e0-94bd-157c6f079c05\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.425617 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hbvt\" (UniqueName: \"kubernetes.io/projected/06177501-5baa-49e0-94bd-157c6f079c05-kube-api-access-9hbvt\") pod \"06177501-5baa-49e0-94bd-157c6f079c05\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.425647 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-scripts\") pod \"06177501-5baa-49e0-94bd-157c6f079c05\" (UID: \"06177501-5baa-49e0-94bd-157c6f079c05\") " Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.426032 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06177501-5baa-49e0-94bd-157c6f079c05-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.432472 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06177501-5baa-49e0-94bd-157c6f079c05-kube-api-access-9hbvt" (OuterVolumeSpecName: "kube-api-access-9hbvt") pod "06177501-5baa-49e0-94bd-157c6f079c05" (UID: "06177501-5baa-49e0-94bd-157c6f079c05"). InnerVolumeSpecName "kube-api-access-9hbvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.432640 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-scripts" (OuterVolumeSpecName: "scripts") pod "06177501-5baa-49e0-94bd-157c6f079c05" (UID: "06177501-5baa-49e0-94bd-157c6f079c05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.432696 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "06177501-5baa-49e0-94bd-157c6f079c05" (UID: "06177501-5baa-49e0-94bd-157c6f079c05"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.458496 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06177501-5baa-49e0-94bd-157c6f079c05" (UID: "06177501-5baa-49e0-94bd-157c6f079c05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.513553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-config-data" (OuterVolumeSpecName: "config-data") pod "06177501-5baa-49e0-94bd-157c6f079c05" (UID: "06177501-5baa-49e0-94bd-157c6f079c05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.527705 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.527739 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.527750 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.527767 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hbvt\" (UniqueName: \"kubernetes.io/projected/06177501-5baa-49e0-94bd-157c6f079c05-kube-api-access-9hbvt\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.527779 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06177501-5baa-49e0-94bd-157c6f079c05-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.730902 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5897b758c4-ntx2t"] Dec 08 15:04:56 crc kubenswrapper[4894]: E1208 15:04:56.731379 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f57102-ceaf-484a-a0b3-c1ecf324339d" containerName="barbican-db-sync" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.731403 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f57102-ceaf-484a-a0b3-c1ecf324339d" containerName="barbican-db-sync" Dec 08 15:04:56 crc kubenswrapper[4894]: E1208 15:04:56.731442 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerName="init" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.731451 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerName="init" Dec 08 15:04:56 crc kubenswrapper[4894]: E1208 15:04:56.731473 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerName="dnsmasq-dns" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.731483 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerName="dnsmasq-dns" Dec 08 15:04:56 crc kubenswrapper[4894]: E1208 15:04:56.731507 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06177501-5baa-49e0-94bd-157c6f079c05" containerName="cinder-db-sync" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.731516 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="06177501-5baa-49e0-94bd-157c6f079c05" containerName="cinder-db-sync" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.731719 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="06177501-5baa-49e0-94bd-157c6f079c05" containerName="cinder-db-sync" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.731743 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f57102-ceaf-484a-a0b3-c1ecf324339d" containerName="barbican-db-sync" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.731754 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97e76d4c-be0b-4ed9-8e51-c5bed27d5514" containerName="dnsmasq-dns" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.733016 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.738264 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.738564 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.738753 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-h78m5" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.770038 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-845d6bc787-swrrt"] Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.771666 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.775398 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.790367 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5897b758c4-ntx2t"] Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.811585 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-845d6bc787-swrrt"] Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.827762 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7nxjg" event={"ID":"06177501-5baa-49e0-94bd-157c6f079c05","Type":"ContainerDied","Data":"45260764b5a80ebb2fd1b7736a8a527c6488e09823c9208279392d2bda1db5ec"} Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.827802 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45260764b5a80ebb2fd1b7736a8a527c6488e09823c9208279392d2bda1db5ec" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.827879 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7nxjg" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.834950 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktb2m\" (UniqueName: \"kubernetes.io/projected/f3b10111-9cb4-4baa-863b-3c34341b2e07-kube-api-access-ktb2m\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835054 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-combined-ca-bundle\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835101 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-config-data-custom\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835132 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-config-data-custom\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835164 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-combined-ca-bundle\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835186 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-config-data\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835205 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b10111-9cb4-4baa-863b-3c34341b2e07-logs\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvr8s\" (UniqueName: \"kubernetes.io/projected/561cc404-97e2-4808-9c11-e83efcb00484-kube-api-access-hvr8s\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835383 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/561cc404-97e2-4808-9c11-e83efcb00484-logs\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.835422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-config-data\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.881947 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-5dfsw"] Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.885031 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.908369 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-5dfsw"] Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938287 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938370 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-config-data-custom\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-config-data-custom\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d4n4\" (UniqueName: \"kubernetes.io/projected/ac053938-f54d-4714-b316-ab808b5f7a2d-kube-api-access-4d4n4\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938536 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938570 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-combined-ca-bundle\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938599 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-config-data\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938632 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b10111-9cb4-4baa-863b-3c34341b2e07-logs\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938662 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938701 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-config\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938744 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvr8s\" (UniqueName: \"kubernetes.io/projected/561cc404-97e2-4808-9c11-e83efcb00484-kube-api-access-hvr8s\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938789 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/561cc404-97e2-4808-9c11-e83efcb00484-logs\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938841 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-config-data\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.938933 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktb2m\" (UniqueName: \"kubernetes.io/projected/f3b10111-9cb4-4baa-863b-3c34341b2e07-kube-api-access-ktb2m\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.939043 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-combined-ca-bundle\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.946440 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-config-data-custom\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.946920 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3b10111-9cb4-4baa-863b-3c34341b2e07-logs\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.947232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/561cc404-97e2-4808-9c11-e83efcb00484-logs\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.947914 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-combined-ca-bundle\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.952172 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-combined-ca-bundle\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.958335 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-config-data\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.960167 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/561cc404-97e2-4808-9c11-e83efcb00484-config-data-custom\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.965939 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3b10111-9cb4-4baa-863b-3c34341b2e07-config-data\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.982260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktb2m\" (UniqueName: \"kubernetes.io/projected/f3b10111-9cb4-4baa-863b-3c34341b2e07-kube-api-access-ktb2m\") pod \"barbican-worker-845d6bc787-swrrt\" (UID: \"f3b10111-9cb4-4baa-863b-3c34341b2e07\") " pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:56 crc kubenswrapper[4894]: I1208 15:04:56.987277 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvr8s\" (UniqueName: \"kubernetes.io/projected/561cc404-97e2-4808-9c11-e83efcb00484-kube-api-access-hvr8s\") pod \"barbican-keystone-listener-5897b758c4-ntx2t\" (UID: \"561cc404-97e2-4808-9c11-e83efcb00484\") " pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.041225 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.041292 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.041326 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d4n4\" (UniqueName: \"kubernetes.io/projected/ac053938-f54d-4714-b316-ab808b5f7a2d-kube-api-access-4d4n4\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.041346 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.041372 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.041391 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-config\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.042330 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-config\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.042898 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.043391 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.044129 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.044648 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.055279 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-86796dcf68-kb8nc"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.058363 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.063117 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.069303 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.082950 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-86796dcf68-kb8nc"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.104360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d4n4\" (UniqueName: \"kubernetes.io/projected/ac053938-f54d-4714-b316-ab808b5f7a2d-kube-api-access-4d4n4\") pod \"dnsmasq-dns-75c8ddd69c-5dfsw\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.125370 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-845d6bc787-swrrt" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.143440 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4rxl\" (UniqueName: \"kubernetes.io/projected/1805e51a-2b46-4c58-890c-8bfe560b74f3-kube-api-access-z4rxl\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.143522 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1805e51a-2b46-4c58-890c-8bfe560b74f3-logs\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.143569 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data-custom\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.143683 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-combined-ca-bundle\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.143731 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.205009 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.231901 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.233617 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.236738 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.245752 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-combined-ca-bundle\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.245839 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.245899 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4rxl\" (UniqueName: \"kubernetes.io/projected/1805e51a-2b46-4c58-890c-8bfe560b74f3-kube-api-access-z4rxl\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.245930 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1805e51a-2b46-4c58-890c-8bfe560b74f3-logs\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.245954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data-custom\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.247171 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.248119 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zjln5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.248289 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.248638 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1805e51a-2b46-4c58-890c-8bfe560b74f3-logs\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.254833 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data-custom\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.259922 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.260469 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-combined-ca-bundle\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.276382 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-5dfsw"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.289123 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4rxl\" (UniqueName: \"kubernetes.io/projected/1805e51a-2b46-4c58-890c-8bfe560b74f3-kube-api-access-z4rxl\") pod \"barbican-api-86796dcf68-kb8nc\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.290804 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.298876 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.326803 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2gpp5"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.328368 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348165 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-scripts\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348232 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmf4n\" (UniqueName: \"kubernetes.io/projected/b6ea0e91-5015-4954-ac06-8dad5fd37190-kube-api-access-lmf4n\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348307 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghhhn\" (UniqueName: \"kubernetes.io/projected/ee3e5d72-93bc-437f-95eb-59ced7cb4590-kube-api-access-ghhhn\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348329 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-config\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348353 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee3e5d72-93bc-437f-95eb-59ced7cb4590-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348381 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348415 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348471 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348487 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348511 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.348529 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.355578 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2gpp5"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.415030 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.451934 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.455300 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.455433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.455568 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.455622 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.455655 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.455698 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.455753 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.456124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-scripts\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.456147 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.456180 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmf4n\" (UniqueName: \"kubernetes.io/projected/b6ea0e91-5015-4954-ac06-8dad5fd37190-kube-api-access-lmf4n\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.456383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghhhn\" (UniqueName: \"kubernetes.io/projected/ee3e5d72-93bc-437f-95eb-59ced7cb4590-kube-api-access-ghhhn\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.456427 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-config\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.456701 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee3e5d72-93bc-437f-95eb-59ced7cb4590-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.456980 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee3e5d72-93bc-437f-95eb-59ced7cb4590-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.458105 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-svc\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.465516 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.466792 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.467493 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.468054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.492213 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-config\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.504281 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.507453 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-scripts\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.511596 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.511735 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.522422 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.568931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmf4n\" (UniqueName: \"kubernetes.io/projected/b6ea0e91-5015-4954-ac06-8dad5fd37190-kube-api-access-lmf4n\") pod \"dnsmasq-dns-5784cf869f-2gpp5\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.570807 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data-custom\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.570866 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9617baee-f4fd-4b22-8526-acc6b07ff690-logs\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.570903 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.570957 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh6wf\" (UniqueName: \"kubernetes.io/projected/9617baee-f4fd-4b22-8526-acc6b07ff690-kube-api-access-mh6wf\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.570979 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.571014 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-scripts\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.571052 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9617baee-f4fd-4b22-8526-acc6b07ff690-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.578854 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghhhn\" (UniqueName: \"kubernetes.io/projected/ee3e5d72-93bc-437f-95eb-59ced7cb4590-kube-api-access-ghhhn\") pod \"cinder-scheduler-0\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.614260 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.671323 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.678626 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data-custom\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.678722 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9617baee-f4fd-4b22-8526-acc6b07ff690-logs\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.678762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.678795 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh6wf\" (UniqueName: \"kubernetes.io/projected/9617baee-f4fd-4b22-8526-acc6b07ff690-kube-api-access-mh6wf\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.678837 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.678867 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-scripts\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.678900 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9617baee-f4fd-4b22-8526-acc6b07ff690-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.679038 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9617baee-f4fd-4b22-8526-acc6b07ff690-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.681258 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9617baee-f4fd-4b22-8526-acc6b07ff690-logs\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.690888 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data-custom\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.693383 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-scripts\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.693923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.698280 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.699155 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh6wf\" (UniqueName: \"kubernetes.io/projected/9617baee-f4fd-4b22-8526-acc6b07ff690-kube-api-access-mh6wf\") pod \"cinder-api-0\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " pod="openstack/cinder-api-0" Dec 08 15:04:57 crc kubenswrapper[4894]: I1208 15:04:57.956567 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 15:04:58 crc kubenswrapper[4894]: I1208 15:04:58.687596 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5897b758c4-ntx2t"] Dec 08 15:04:58 crc kubenswrapper[4894]: I1208 15:04:58.760050 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-5dfsw"] Dec 08 15:04:58 crc kubenswrapper[4894]: I1208 15:04:58.823156 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:04:58 crc kubenswrapper[4894]: I1208 15:04:58.868090 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" event={"ID":"ac053938-f54d-4714-b316-ab808b5f7a2d","Type":"ContainerStarted","Data":"9400b6b31892058b7b41e75ceb6760a2879955117133897cb484d3fdde8a55f3"} Dec 08 15:04:58 crc kubenswrapper[4894]: I1208 15:04:58.874260 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" event={"ID":"561cc404-97e2-4808-9c11-e83efcb00484","Type":"ContainerStarted","Data":"e427d1f26625a41a0aa7037b9cfd69d91235a8685f8e0192080d3554bd19c3e0"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.080209 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-86796dcf68-kb8nc"] Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.095841 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-845d6bc787-swrrt"] Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.200009 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.212868 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2gpp5"] Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.727127 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.890827 4894 generic.go:334] "Generic (PLEG): container finished" podID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerID="bc20d2242126f668c23a26db8565128f10768432421c56abe86521cf8c071199" exitCode=0 Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.891061 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" event={"ID":"b6ea0e91-5015-4954-ac06-8dad5fd37190","Type":"ContainerDied","Data":"bc20d2242126f668c23a26db8565128f10768432421c56abe86521cf8c071199"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.891228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" event={"ID":"b6ea0e91-5015-4954-ac06-8dad5fd37190","Type":"ContainerStarted","Data":"51b48f356c1cf26715b90d4d3ffcc43efca56e233c0809a116c7a59d9b8c371b"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.901652 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee3e5d72-93bc-437f-95eb-59ced7cb4590","Type":"ContainerStarted","Data":"739ac44faecfa962fdc5c5e632d561be7ea1839616d41655d390bc9ae924a543"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.909226 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845d6bc787-swrrt" event={"ID":"f3b10111-9cb4-4baa-863b-3c34341b2e07","Type":"ContainerStarted","Data":"708b3a885c291822e3251ebc0540ce0f958dea1a38ab8dcca2f7bbea41646c5a"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.927616 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86796dcf68-kb8nc" event={"ID":"1805e51a-2b46-4c58-890c-8bfe560b74f3","Type":"ContainerStarted","Data":"95e118df7c94651c2914923cdf75d03fa41f501e562cabef577a60f94a0ee1ab"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.927708 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86796dcf68-kb8nc" event={"ID":"1805e51a-2b46-4c58-890c-8bfe560b74f3","Type":"ContainerStarted","Data":"f1d1c24cc60914a0213b76ff2b12244b7908f3edb4f8aae3c27c5b1365c9bb94"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.928041 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.928077 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86796dcf68-kb8nc" event={"ID":"1805e51a-2b46-4c58-890c-8bfe560b74f3","Type":"ContainerStarted","Data":"19144d8a407ae983e1b1df1ee98c6c03634257a15a97bf0c44cee1817aee3479"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.928097 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.931883 4894 generic.go:334] "Generic (PLEG): container finished" podID="ac053938-f54d-4714-b316-ab808b5f7a2d" containerID="e3452f0a9e7d0db4ef898885ffadad684b50e329819d3cb90dd9d5f00bcf4f81" exitCode=0 Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.931952 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" event={"ID":"ac053938-f54d-4714-b316-ab808b5f7a2d","Type":"ContainerDied","Data":"e3452f0a9e7d0db4ef898885ffadad684b50e329819d3cb90dd9d5f00bcf4f81"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.935938 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9617baee-f4fd-4b22-8526-acc6b07ff690","Type":"ContainerStarted","Data":"1f0d621ac2ac1480aebd58d416219f7c4c4c0a24a2cb6acc16b488e08bafe3e5"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.941341 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerStarted","Data":"a0783f170bd1734cdc76c747073ceab22b9e701d7027c11711dfea2e1f57af2b"} Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.941582 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-central-agent" containerID="cri-o://a258a59f2c327d7149b40ebefb52a70112b73cda3bd4d7b858fc52d4154f5ee8" gracePeriod=30 Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.942143 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.942225 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="proxy-httpd" containerID="cri-o://a0783f170bd1734cdc76c747073ceab22b9e701d7027c11711dfea2e1f57af2b" gracePeriod=30 Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.942296 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="sg-core" containerID="cri-o://d5a83a8baf3ba09d4e59a5447a2b742b13cb1f8293257083f73de6104b9d931b" gracePeriod=30 Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.942391 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-notification-agent" containerID="cri-o://1af977c38788e3c71dbb4eee715a307b85476868c54ad8969918772446207d0e" gracePeriod=30 Dec 08 15:04:59 crc kubenswrapper[4894]: I1208 15:04:59.951971 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-86796dcf68-kb8nc" podStartSLOduration=2.951952517 podStartE2EDuration="2.951952517s" podCreationTimestamp="2025-12-08 15:04:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:04:59.950607615 +0000 UTC m=+1121.050613750" watchObservedRunningTime="2025-12-08 15:04:59.951952517 +0000 UTC m=+1121.051958632" Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.037071 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.337556035 podStartE2EDuration="1m0.037050602s" podCreationTimestamp="2025-12-08 15:04:00 +0000 UTC" firstStartedPulling="2025-12-08 15:04:02.472236745 +0000 UTC m=+1063.572242850" lastFinishedPulling="2025-12-08 15:04:58.171731302 +0000 UTC m=+1119.271737417" observedRunningTime="2025-12-08 15:04:59.986250931 +0000 UTC m=+1121.086257046" watchObservedRunningTime="2025-12-08 15:05:00.037050602 +0000 UTC m=+1121.137056717" Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.588963 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.966687 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9617baee-f4fd-4b22-8526-acc6b07ff690","Type":"ContainerStarted","Data":"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1"} Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.969529 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" event={"ID":"ac053938-f54d-4714-b316-ab808b5f7a2d","Type":"ContainerDied","Data":"9400b6b31892058b7b41e75ceb6760a2879955117133897cb484d3fdde8a55f3"} Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.969589 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9400b6b31892058b7b41e75ceb6760a2879955117133897cb484d3fdde8a55f3" Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.987067 4894 generic.go:334] "Generic (PLEG): container finished" podID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerID="a0783f170bd1734cdc76c747073ceab22b9e701d7027c11711dfea2e1f57af2b" exitCode=0 Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.987134 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerDied","Data":"a0783f170bd1734cdc76c747073ceab22b9e701d7027c11711dfea2e1f57af2b"} Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.987835 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerDied","Data":"d5a83a8baf3ba09d4e59a5447a2b742b13cb1f8293257083f73de6104b9d931b"} Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.988007 4894 generic.go:334] "Generic (PLEG): container finished" podID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerID="d5a83a8baf3ba09d4e59a5447a2b742b13cb1f8293257083f73de6104b9d931b" exitCode=2 Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.988082 4894 generic.go:334] "Generic (PLEG): container finished" podID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerID="1af977c38788e3c71dbb4eee715a307b85476868c54ad8969918772446207d0e" exitCode=0 Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.988238 4894 generic.go:334] "Generic (PLEG): container finished" podID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerID="a258a59f2c327d7149b40ebefb52a70112b73cda3bd4d7b858fc52d4154f5ee8" exitCode=0 Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.988319 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerDied","Data":"1af977c38788e3c71dbb4eee715a307b85476868c54ad8969918772446207d0e"} Dec 08 15:05:00 crc kubenswrapper[4894]: I1208 15:05:00.990065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerDied","Data":"a258a59f2c327d7149b40ebefb52a70112b73cda3bd4d7b858fc52d4154f5ee8"} Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.047854 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.202041 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-sb\") pod \"ac053938-f54d-4714-b316-ab808b5f7a2d\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.202451 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-swift-storage-0\") pod \"ac053938-f54d-4714-b316-ab808b5f7a2d\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.202481 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-config\") pod \"ac053938-f54d-4714-b316-ab808b5f7a2d\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.202549 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-nb\") pod \"ac053938-f54d-4714-b316-ab808b5f7a2d\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.202578 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-svc\") pod \"ac053938-f54d-4714-b316-ab808b5f7a2d\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.202760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4n4\" (UniqueName: \"kubernetes.io/projected/ac053938-f54d-4714-b316-ab808b5f7a2d-kube-api-access-4d4n4\") pod \"ac053938-f54d-4714-b316-ab808b5f7a2d\" (UID: \"ac053938-f54d-4714-b316-ab808b5f7a2d\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.235614 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac053938-f54d-4714-b316-ab808b5f7a2d-kube-api-access-4d4n4" (OuterVolumeSpecName: "kube-api-access-4d4n4") pod "ac053938-f54d-4714-b316-ab808b5f7a2d" (UID: "ac053938-f54d-4714-b316-ab808b5f7a2d"). InnerVolumeSpecName "kube-api-access-4d4n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.238062 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac053938-f54d-4714-b316-ab808b5f7a2d" (UID: "ac053938-f54d-4714-b316-ab808b5f7a2d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.238559 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ac053938-f54d-4714-b316-ab808b5f7a2d" (UID: "ac053938-f54d-4714-b316-ab808b5f7a2d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.238833 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ac053938-f54d-4714-b316-ab808b5f7a2d" (UID: "ac053938-f54d-4714-b316-ab808b5f7a2d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.247292 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-config" (OuterVolumeSpecName: "config") pod "ac053938-f54d-4714-b316-ab808b5f7a2d" (UID: "ac053938-f54d-4714-b316-ab808b5f7a2d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.248782 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ac053938-f54d-4714-b316-ab808b5f7a2d" (UID: "ac053938-f54d-4714-b316-ab808b5f7a2d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.305440 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4n4\" (UniqueName: \"kubernetes.io/projected/ac053938-f54d-4714-b316-ab808b5f7a2d-kube-api-access-4d4n4\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.305481 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.305495 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.305681 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.306208 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.306235 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac053938-f54d-4714-b316-ab808b5f7a2d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.485204 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.612394 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-combined-ca-bundle\") pod \"72ff9384-4eab-4600-a3e7-448b15471c6e\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.612482 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-run-httpd\") pod \"72ff9384-4eab-4600-a3e7-448b15471c6e\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.612693 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-scripts\") pod \"72ff9384-4eab-4600-a3e7-448b15471c6e\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.612735 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-log-httpd\") pod \"72ff9384-4eab-4600-a3e7-448b15471c6e\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.612785 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-config-data\") pod \"72ff9384-4eab-4600-a3e7-448b15471c6e\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.612996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-sg-core-conf-yaml\") pod \"72ff9384-4eab-4600-a3e7-448b15471c6e\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.613237 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prfk6\" (UniqueName: \"kubernetes.io/projected/72ff9384-4eab-4600-a3e7-448b15471c6e-kube-api-access-prfk6\") pod \"72ff9384-4eab-4600-a3e7-448b15471c6e\" (UID: \"72ff9384-4eab-4600-a3e7-448b15471c6e\") " Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.613924 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "72ff9384-4eab-4600-a3e7-448b15471c6e" (UID: "72ff9384-4eab-4600-a3e7-448b15471c6e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.614454 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.614643 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "72ff9384-4eab-4600-a3e7-448b15471c6e" (UID: "72ff9384-4eab-4600-a3e7-448b15471c6e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.627998 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-scripts" (OuterVolumeSpecName: "scripts") pod "72ff9384-4eab-4600-a3e7-448b15471c6e" (UID: "72ff9384-4eab-4600-a3e7-448b15471c6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.628050 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ff9384-4eab-4600-a3e7-448b15471c6e-kube-api-access-prfk6" (OuterVolumeSpecName: "kube-api-access-prfk6") pod "72ff9384-4eab-4600-a3e7-448b15471c6e" (UID: "72ff9384-4eab-4600-a3e7-448b15471c6e"). InnerVolumeSpecName "kube-api-access-prfk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.660727 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "72ff9384-4eab-4600-a3e7-448b15471c6e" (UID: "72ff9384-4eab-4600-a3e7-448b15471c6e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.716739 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.716802 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72ff9384-4eab-4600-a3e7-448b15471c6e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.716842 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.716858 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prfk6\" (UniqueName: \"kubernetes.io/projected/72ff9384-4eab-4600-a3e7-448b15471c6e-kube-api-access-prfk6\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.795003 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72ff9384-4eab-4600-a3e7-448b15471c6e" (UID: "72ff9384-4eab-4600-a3e7-448b15471c6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.810799 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-config-data" (OuterVolumeSpecName: "config-data") pod "72ff9384-4eab-4600-a3e7-448b15471c6e" (UID: "72ff9384-4eab-4600-a3e7-448b15471c6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.818916 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:01 crc kubenswrapper[4894]: I1208 15:05:01.818966 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72ff9384-4eab-4600-a3e7-448b15471c6e-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.055313 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.055368 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72ff9384-4eab-4600-a3e7-448b15471c6e","Type":"ContainerDied","Data":"1cc9664f3dcbe84bf0e0a37cc39b072eb13a4f8025b937080fbde8d36438b07b"} Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.055462 4894 scope.go:117] "RemoveContainer" containerID="a0783f170bd1734cdc76c747073ceab22b9e701d7027c11711dfea2e1f57af2b" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.065586 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" event={"ID":"b6ea0e91-5015-4954-ac06-8dad5fd37190","Type":"ContainerStarted","Data":"c56caa9679e2e7ffd69bb6faf7336d6210d35a05890dcd96203884cbdc6cc5bb"} Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.065693 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.083973 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845d6bc787-swrrt" event={"ID":"f3b10111-9cb4-4baa-863b-3c34341b2e07","Type":"ContainerStarted","Data":"25d518e81b57986bb0c836644b20863d275191dcdbf3f41d7ce9d2af665a47df"} Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.102105 4894 scope.go:117] "RemoveContainer" containerID="d5a83a8baf3ba09d4e59a5447a2b742b13cb1f8293257083f73de6104b9d931b" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.106196 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" podStartSLOduration=5.106110173 podStartE2EDuration="5.106110173s" podCreationTimestamp="2025-12-08 15:04:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:02.090531325 +0000 UTC m=+1123.190537440" watchObservedRunningTime="2025-12-08 15:05:02.106110173 +0000 UTC m=+1123.206116288" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.111258 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-5dfsw" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.114454 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" event={"ID":"561cc404-97e2-4808-9c11-e83efcb00484","Type":"ContainerStarted","Data":"b5f6cdd5ac5c092996138da11b37f4c05bc1b5b52476a2dbb2c925ea8bdd6935"} Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.123830 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.157316 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.211095 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:02 crc kubenswrapper[4894]: E1208 15:05:02.212658 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-notification-agent" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.212688 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-notification-agent" Dec 08 15:05:02 crc kubenswrapper[4894]: E1208 15:05:02.212735 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="sg-core" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.212744 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="sg-core" Dec 08 15:05:02 crc kubenswrapper[4894]: E1208 15:05:02.212785 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="proxy-httpd" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.212793 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="proxy-httpd" Dec 08 15:05:02 crc kubenswrapper[4894]: E1208 15:05:02.212808 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-central-agent" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.212838 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-central-agent" Dec 08 15:05:02 crc kubenswrapper[4894]: E1208 15:05:02.212850 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac053938-f54d-4714-b316-ab808b5f7a2d" containerName="init" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.212858 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac053938-f54d-4714-b316-ab808b5f7a2d" containerName="init" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.213373 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="sg-core" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.213423 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="proxy-httpd" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.213440 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac053938-f54d-4714-b316-ab808b5f7a2d" containerName="init" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.213451 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-central-agent" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.213471 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" containerName="ceilometer-notification-agent" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.224844 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.233711 4894 scope.go:117] "RemoveContainer" containerID="1af977c38788e3c71dbb4eee715a307b85476868c54ad8969918772446207d0e" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.236437 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.236748 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.238500 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.284682 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-5dfsw"] Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.294447 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-5dfsw"] Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.313083 4894 scope.go:117] "RemoveContainer" containerID="a258a59f2c327d7149b40ebefb52a70112b73cda3bd4d7b858fc52d4154f5ee8" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.367480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-scripts\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.367552 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-config-data\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.367585 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.367670 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x74d\" (UniqueName: \"kubernetes.io/projected/f25b4581-989b-4f47-b680-063ea7eb57f7-kube-api-access-6x74d\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.367710 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-run-httpd\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.367755 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-log-httpd\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.367788 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.469000 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.469445 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x74d\" (UniqueName: \"kubernetes.io/projected/f25b4581-989b-4f47-b680-063ea7eb57f7-kube-api-access-6x74d\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.469478 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-run-httpd\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.469509 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-log-httpd\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.469535 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.469639 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-scripts\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.469663 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-config-data\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.470242 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-log-httpd\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.474157 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-run-httpd\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.481933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.483269 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-config-data\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.485265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-scripts\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.486262 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.486532 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x74d\" (UniqueName: \"kubernetes.io/projected/f25b4581-989b-4f47-b680-063ea7eb57f7-kube-api-access-6x74d\") pod \"ceilometer-0\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.564372 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.614683 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:05:02 crc kubenswrapper[4894]: I1208 15:05:02.757192 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.128389 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9617baee-f4fd-4b22-8526-acc6b07ff690","Type":"ContainerStarted","Data":"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0"} Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.128560 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api-log" containerID="cri-o://e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1" gracePeriod=30 Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.128578 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api" containerID="cri-o://7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0" gracePeriod=30 Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.128983 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.136662 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee3e5d72-93bc-437f-95eb-59ced7cb4590","Type":"ContainerStarted","Data":"8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535"} Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.142449 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-845d6bc787-swrrt" event={"ID":"f3b10111-9cb4-4baa-863b-3c34341b2e07","Type":"ContainerStarted","Data":"7ee5a6cb0d37ba1d2fa2ab4bd595e2389634e26d438bbb2c86029dcbba0e4697"} Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.160012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" event={"ID":"561cc404-97e2-4808-9c11-e83efcb00484","Type":"ContainerStarted","Data":"85cf9387573849d088641aa6ffc2d908289cc9e0c521efa50f58dfd383b74b6f"} Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.160803 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.167399 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.167366171 podStartE2EDuration="6.167366171s" podCreationTimestamp="2025-12-08 15:04:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:03.150549664 +0000 UTC m=+1124.250555779" watchObservedRunningTime="2025-12-08 15:05:03.167366171 +0000 UTC m=+1124.267372286" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.175406 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-845d6bc787-swrrt" podStartSLOduration=4.763606978 podStartE2EDuration="7.175381041s" podCreationTimestamp="2025-12-08 15:04:56 +0000 UTC" firstStartedPulling="2025-12-08 15:04:59.099252272 +0000 UTC m=+1120.199258387" lastFinishedPulling="2025-12-08 15:05:01.511026345 +0000 UTC m=+1122.611032450" observedRunningTime="2025-12-08 15:05:03.168805715 +0000 UTC m=+1124.268811840" watchObservedRunningTime="2025-12-08 15:05:03.175381041 +0000 UTC m=+1124.275387156" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.213732 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5897b758c4-ntx2t" podStartSLOduration=4.456656604 podStartE2EDuration="7.213701081s" podCreationTimestamp="2025-12-08 15:04:56 +0000 UTC" firstStartedPulling="2025-12-08 15:04:58.719191229 +0000 UTC m=+1119.819197344" lastFinishedPulling="2025-12-08 15:05:01.476235706 +0000 UTC m=+1122.576241821" observedRunningTime="2025-12-08 15:05:03.192021343 +0000 UTC m=+1124.292027458" watchObservedRunningTime="2025-12-08 15:05:03.213701081 +0000 UTC m=+1124.313707196" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.256229 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72ff9384-4eab-4600-a3e7-448b15471c6e" path="/var/lib/kubelet/pods/72ff9384-4eab-4600-a3e7-448b15471c6e/volumes" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.258554 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac053938-f54d-4714-b316-ab808b5f7a2d" path="/var/lib/kubelet/pods/ac053938-f54d-4714-b316-ab808b5f7a2d/volumes" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.834591 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-789cf67ffd-s9792"] Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.837600 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.867123 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.871113 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.882943 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-789cf67ffd-s9792"] Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.934650 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-config-data\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.934709 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-internal-tls-certs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.934740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwvjz\" (UniqueName: \"kubernetes.io/projected/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-kube-api-access-lwvjz\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.934773 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-public-tls-certs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.934805 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-combined-ca-bundle\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.934858 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-logs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:03 crc kubenswrapper[4894]: I1208 15:05:03.934876 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-config-data-custom\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.027480 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.039013 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-config-data\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.039075 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-internal-tls-certs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.039101 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwvjz\" (UniqueName: \"kubernetes.io/projected/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-kube-api-access-lwvjz\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.039134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-public-tls-certs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.039166 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-combined-ca-bundle\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.039204 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-logs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.039222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-config-data-custom\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.050573 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-internal-tls-certs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.050630 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-config-data-custom\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.051030 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-logs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.061121 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-combined-ca-bundle\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.061838 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-config-data\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.063763 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-public-tls-certs\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.085360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwvjz\" (UniqueName: \"kubernetes.io/projected/7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5-kube-api-access-lwvjz\") pod \"barbican-api-789cf67ffd-s9792\" (UID: \"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5\") " pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.142564 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh6wf\" (UniqueName: \"kubernetes.io/projected/9617baee-f4fd-4b22-8526-acc6b07ff690-kube-api-access-mh6wf\") pod \"9617baee-f4fd-4b22-8526-acc6b07ff690\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.143029 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-combined-ca-bundle\") pod \"9617baee-f4fd-4b22-8526-acc6b07ff690\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.143098 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9617baee-f4fd-4b22-8526-acc6b07ff690-logs\") pod \"9617baee-f4fd-4b22-8526-acc6b07ff690\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.143154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data-custom\") pod \"9617baee-f4fd-4b22-8526-acc6b07ff690\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.143224 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data\") pod \"9617baee-f4fd-4b22-8526-acc6b07ff690\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.143287 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9617baee-f4fd-4b22-8526-acc6b07ff690-etc-machine-id\") pod \"9617baee-f4fd-4b22-8526-acc6b07ff690\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.143325 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-scripts\") pod \"9617baee-f4fd-4b22-8526-acc6b07ff690\" (UID: \"9617baee-f4fd-4b22-8526-acc6b07ff690\") " Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.146195 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9617baee-f4fd-4b22-8526-acc6b07ff690-logs" (OuterVolumeSpecName: "logs") pod "9617baee-f4fd-4b22-8526-acc6b07ff690" (UID: "9617baee-f4fd-4b22-8526-acc6b07ff690"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.146729 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9617baee-f4fd-4b22-8526-acc6b07ff690-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9617baee-f4fd-4b22-8526-acc6b07ff690" (UID: "9617baee-f4fd-4b22-8526-acc6b07ff690"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.151577 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9617baee-f4fd-4b22-8526-acc6b07ff690-kube-api-access-mh6wf" (OuterVolumeSpecName: "kube-api-access-mh6wf") pod "9617baee-f4fd-4b22-8526-acc6b07ff690" (UID: "9617baee-f4fd-4b22-8526-acc6b07ff690"). InnerVolumeSpecName "kube-api-access-mh6wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.174773 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9617baee-f4fd-4b22-8526-acc6b07ff690" (UID: "9617baee-f4fd-4b22-8526-acc6b07ff690"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.175425 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-scripts" (OuterVolumeSpecName: "scripts") pod "9617baee-f4fd-4b22-8526-acc6b07ff690" (UID: "9617baee-f4fd-4b22-8526-acc6b07ff690"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.231631 4894 generic.go:334] "Generic (PLEG): container finished" podID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerID="7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0" exitCode=0 Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.231715 4894 generic.go:334] "Generic (PLEG): container finished" podID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerID="e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1" exitCode=143 Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.231900 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9617baee-f4fd-4b22-8526-acc6b07ff690","Type":"ContainerDied","Data":"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0"} Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.231951 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9617baee-f4fd-4b22-8526-acc6b07ff690","Type":"ContainerDied","Data":"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1"} Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.231968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9617baee-f4fd-4b22-8526-acc6b07ff690","Type":"ContainerDied","Data":"1f0d621ac2ac1480aebd58d416219f7c4c4c0a24a2cb6acc16b488e08bafe3e5"} Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.231992 4894 scope.go:117] "RemoveContainer" containerID="7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.232623 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.245368 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9617baee-f4fd-4b22-8526-acc6b07ff690-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.245414 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.245426 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh6wf\" (UniqueName: \"kubernetes.io/projected/9617baee-f4fd-4b22-8526-acc6b07ff690-kube-api-access-mh6wf\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.245440 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9617baee-f4fd-4b22-8526-acc6b07ff690-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.245454 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.256263 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.258006 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9617baee-f4fd-4b22-8526-acc6b07ff690" (UID: "9617baee-f4fd-4b22-8526-acc6b07ff690"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.270076 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee3e5d72-93bc-437f-95eb-59ced7cb4590","Type":"ContainerStarted","Data":"402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d"} Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.274864 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data" (OuterVolumeSpecName: "config-data") pod "9617baee-f4fd-4b22-8526-acc6b07ff690" (UID: "9617baee-f4fd-4b22-8526-acc6b07ff690"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.279413 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerStarted","Data":"9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369"} Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.279478 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerStarted","Data":"fe7e26deba89966a0467f3ad00ea49c4ed6c3b0cf2e828a2a2a5916feda1213a"} Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.296502 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.297444453 podStartE2EDuration="7.296483042s" podCreationTimestamp="2025-12-08 15:04:57 +0000 UTC" firstStartedPulling="2025-12-08 15:04:58.860356209 +0000 UTC m=+1119.960362324" lastFinishedPulling="2025-12-08 15:04:59.859394798 +0000 UTC m=+1120.959400913" observedRunningTime="2025-12-08 15:05:04.295720459 +0000 UTC m=+1125.395726584" watchObservedRunningTime="2025-12-08 15:05:04.296483042 +0000 UTC m=+1125.396489157" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.337096 4894 scope.go:117] "RemoveContainer" containerID="e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.348831 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.349856 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9617baee-f4fd-4b22-8526-acc6b07ff690-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.368346 4894 scope.go:117] "RemoveContainer" containerID="7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0" Dec 08 15:05:04 crc kubenswrapper[4894]: E1208 15:05:04.369011 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0\": container with ID starting with 7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0 not found: ID does not exist" containerID="7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.369089 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0"} err="failed to get container status \"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0\": rpc error: code = NotFound desc = could not find container \"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0\": container with ID starting with 7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0 not found: ID does not exist" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.369136 4894 scope.go:117] "RemoveContainer" containerID="e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1" Dec 08 15:05:04 crc kubenswrapper[4894]: E1208 15:05:04.369592 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1\": container with ID starting with e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1 not found: ID does not exist" containerID="e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.369650 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1"} err="failed to get container status \"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1\": rpc error: code = NotFound desc = could not find container \"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1\": container with ID starting with e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1 not found: ID does not exist" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.369683 4894 scope.go:117] "RemoveContainer" containerID="7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.370035 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0"} err="failed to get container status \"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0\": rpc error: code = NotFound desc = could not find container \"7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0\": container with ID starting with 7116fc065c95a8d87ff2c2fb01d33bd4fb4da9c2ae5d80512de8287ec46161d0 not found: ID does not exist" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.370063 4894 scope.go:117] "RemoveContainer" containerID="e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.370346 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1"} err="failed to get container status \"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1\": rpc error: code = NotFound desc = could not find container \"e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1\": container with ID starting with e3e63892cef0245dc6062fb02f154dbad38008f7bcb7e5b635148c6fd897ebd1 not found: ID does not exist" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.609265 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.622463 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.645049 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:05:04 crc kubenswrapper[4894]: E1208 15:05:04.645487 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api-log" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.645506 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api-log" Dec 08 15:05:04 crc kubenswrapper[4894]: E1208 15:05:04.645531 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.645543 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.645755 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.645781 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" containerName="cinder-api-log" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.650138 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.656439 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.656563 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.660243 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.672016 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762123 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762451 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762497 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-config-data\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762520 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-scripts\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762539 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjphs\" (UniqueName: \"kubernetes.io/projected/11765aa0-3b20-4d49-9017-8a56603c10b0-kube-api-access-vjphs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762602 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11765aa0-3b20-4d49-9017-8a56603c10b0-logs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762637 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762709 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11765aa0-3b20-4d49-9017-8a56603c10b0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.762782 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-config-data-custom\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.842131 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-789cf67ffd-s9792"] Dec 08 15:05:04 crc kubenswrapper[4894]: W1208 15:05:04.922822 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c1c1c34_52bc_4425_9f7d_75ef02f6e5a5.slice/crio-76954ab0c5973459d521b97d0cbec065194c8be714fcbd31efd843aed482abb2 WatchSource:0}: Error finding container 76954ab0c5973459d521b97d0cbec065194c8be714fcbd31efd843aed482abb2: Status 404 returned error can't find the container with id 76954ab0c5973459d521b97d0cbec065194c8be714fcbd31efd843aed482abb2 Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.927680 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.927854 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11765aa0-3b20-4d49-9017-8a56603c10b0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.927969 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-config-data-custom\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.928036 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.928072 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.928133 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-config-data\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.928175 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-scripts\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.928197 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjphs\" (UniqueName: \"kubernetes.io/projected/11765aa0-3b20-4d49-9017-8a56603c10b0-kube-api-access-vjphs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.928281 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11765aa0-3b20-4d49-9017-8a56603c10b0-logs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.931603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11765aa0-3b20-4d49-9017-8a56603c10b0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.932874 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11765aa0-3b20-4d49-9017-8a56603c10b0-logs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.941374 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-config-data-custom\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.959403 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjphs\" (UniqueName: \"kubernetes.io/projected/11765aa0-3b20-4d49-9017-8a56603c10b0-kube-api-access-vjphs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.964857 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.965762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-scripts\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.966215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-config-data\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.969097 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.979639 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11765aa0-3b20-4d49-9017-8a56603c10b0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"11765aa0-3b20-4d49-9017-8a56603c10b0\") " pod="openstack/cinder-api-0" Dec 08 15:05:04 crc kubenswrapper[4894]: I1208 15:05:04.987347 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.218658 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9617baee-f4fd-4b22-8526-acc6b07ff690" path="/var/lib/kubelet/pods/9617baee-f4fd-4b22-8526-acc6b07ff690/volumes" Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.293329 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerStarted","Data":"61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b"} Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.294543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-789cf67ffd-s9792" event={"ID":"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5","Type":"ContainerStarted","Data":"3794f029566272155218a9c1e6aa904c6db6f6393c4a783d286235c0ad3e0532"} Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.294563 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-789cf67ffd-s9792" event={"ID":"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5","Type":"ContainerStarted","Data":"76954ab0c5973459d521b97d0cbec065194c8be714fcbd31efd843aed482abb2"} Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.509287 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.777236 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.793343 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-797dcfbfb8-294k4" Dec 08 15:05:05 crc kubenswrapper[4894]: I1208 15:05:05.949341 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.021941 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fb57bf44-rk2jn"] Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.350080 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"11765aa0-3b20-4d49-9017-8a56603c10b0","Type":"ContainerStarted","Data":"cd7e4b50e6f057cec2478fe91b904d6f8ca914ab79a9df8124de661aa062a314"} Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.381039 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerStarted","Data":"7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21"} Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.386731 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7fb57bf44-rk2jn" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon-log" containerID="cri-o://49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7" gracePeriod=30 Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.386976 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-789cf67ffd-s9792" event={"ID":"7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5","Type":"ContainerStarted","Data":"97453baae83bf857905051765cd56c82f410b502c6c719c336ac5c22541e5dc7"} Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.387055 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7fb57bf44-rk2jn" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" containerID="cri-o://31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844" gracePeriod=30 Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.387361 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.387395 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:06 crc kubenswrapper[4894]: I1208 15:05:06.422011 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-789cf67ffd-s9792" podStartSLOduration=3.42197264 podStartE2EDuration="3.42197264s" podCreationTimestamp="2025-12-08 15:05:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:06.418436159 +0000 UTC m=+1127.518442274" watchObservedRunningTime="2025-12-08 15:05:06.42197264 +0000 UTC m=+1127.521978755" Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.296659 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.296988 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.398290 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerStarted","Data":"2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f"} Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.398711 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.400659 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"11765aa0-3b20-4d49-9017-8a56603c10b0","Type":"ContainerStarted","Data":"ba643129cc44e4b851f84dbcdad7902e9f000daf1b71b1af9aae5cd4928644bc"} Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.400809 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"11765aa0-3b20-4d49-9017-8a56603c10b0","Type":"ContainerStarted","Data":"f255e3f2ffaea0ebc6989f2e691b4ac3df65cce34383e72320a5d906db45db2d"} Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.428444 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.590412621 podStartE2EDuration="5.428425182s" podCreationTimestamp="2025-12-08 15:05:02 +0000 UTC" firstStartedPulling="2025-12-08 15:05:03.198490626 +0000 UTC m=+1124.298496741" lastFinishedPulling="2025-12-08 15:05:07.036503187 +0000 UTC m=+1128.136509302" observedRunningTime="2025-12-08 15:05:07.420064 +0000 UTC m=+1128.520070115" watchObservedRunningTime="2025-12-08 15:05:07.428425182 +0000 UTC m=+1128.528431297" Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.452123 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.452096813 podStartE2EDuration="3.452096813s" podCreationTimestamp="2025-12-08 15:05:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:07.445396793 +0000 UTC m=+1128.545402908" watchObservedRunningTime="2025-12-08 15:05:07.452096813 +0000 UTC m=+1128.552102928" Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.615073 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.674003 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.784777 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-8sqvg"] Dec 08 15:05:07 crc kubenswrapper[4894]: I1208 15:05:07.785111 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" podUID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerName="dnsmasq-dns" containerID="cri-o://009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147" gracePeriod=10 Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.231249 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.404984 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.410500 4894 generic.go:334] "Generic (PLEG): container finished" podID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerID="009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147" exitCode=0 Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.410587 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.410745 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" event={"ID":"8be96a69-9240-4484-96d0-8b5eb9e44ead","Type":"ContainerDied","Data":"009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147"} Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.410800 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-8sqvg" event={"ID":"8be96a69-9240-4484-96d0-8b5eb9e44ead","Type":"ContainerDied","Data":"8df4570b02e894d918705cb724fe3ce5a4e2ddf9aae23f36dcfa2395c80cf8d3"} Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.410834 4894 scope.go:117] "RemoveContainer" containerID="009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.411376 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.433281 4894 scope.go:117] "RemoveContainer" containerID="fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.471436 4894 scope.go:117] "RemoveContainer" containerID="009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147" Dec 08 15:05:08 crc kubenswrapper[4894]: E1208 15:05:08.472249 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147\": container with ID starting with 009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147 not found: ID does not exist" containerID="009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.472284 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147"} err="failed to get container status \"009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147\": rpc error: code = NotFound desc = could not find container \"009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147\": container with ID starting with 009ab243b754751ff8cc6397f992a71b1189736acd1be7b56fb117898b3e6147 not found: ID does not exist" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.472308 4894 scope.go:117] "RemoveContainer" containerID="fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9" Dec 08 15:05:08 crc kubenswrapper[4894]: E1208 15:05:08.473022 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9\": container with ID starting with fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9 not found: ID does not exist" containerID="fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.473046 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9"} err="failed to get container status \"fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9\": rpc error: code = NotFound desc = could not find container \"fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9\": container with ID starting with fcac5adc9898d60021088711184bd4781eba9569286c8a56c9d6147522556be9 not found: ID does not exist" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.515734 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.554801 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-config\") pod \"8be96a69-9240-4484-96d0-8b5eb9e44ead\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.554935 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-nb\") pod \"8be96a69-9240-4484-96d0-8b5eb9e44ead\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.555078 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-svc\") pod \"8be96a69-9240-4484-96d0-8b5eb9e44ead\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.555102 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpg6h\" (UniqueName: \"kubernetes.io/projected/8be96a69-9240-4484-96d0-8b5eb9e44ead-kube-api-access-gpg6h\") pod \"8be96a69-9240-4484-96d0-8b5eb9e44ead\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.555160 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-swift-storage-0\") pod \"8be96a69-9240-4484-96d0-8b5eb9e44ead\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.555201 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-sb\") pod \"8be96a69-9240-4484-96d0-8b5eb9e44ead\" (UID: \"8be96a69-9240-4484-96d0-8b5eb9e44ead\") " Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.561623 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8be96a69-9240-4484-96d0-8b5eb9e44ead-kube-api-access-gpg6h" (OuterVolumeSpecName: "kube-api-access-gpg6h") pod "8be96a69-9240-4484-96d0-8b5eb9e44ead" (UID: "8be96a69-9240-4484-96d0-8b5eb9e44ead"). InnerVolumeSpecName "kube-api-access-gpg6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.648000 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8be96a69-9240-4484-96d0-8b5eb9e44ead" (UID: "8be96a69-9240-4484-96d0-8b5eb9e44ead"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.648989 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8be96a69-9240-4484-96d0-8b5eb9e44ead" (UID: "8be96a69-9240-4484-96d0-8b5eb9e44ead"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.658595 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.658636 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpg6h\" (UniqueName: \"kubernetes.io/projected/8be96a69-9240-4484-96d0-8b5eb9e44ead-kube-api-access-gpg6h\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.658652 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.659467 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8be96a69-9240-4484-96d0-8b5eb9e44ead" (UID: "8be96a69-9240-4484-96d0-8b5eb9e44ead"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.682435 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-config" (OuterVolumeSpecName: "config") pod "8be96a69-9240-4484-96d0-8b5eb9e44ead" (UID: "8be96a69-9240-4484-96d0-8b5eb9e44ead"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.684351 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8be96a69-9240-4484-96d0-8b5eb9e44ead" (UID: "8be96a69-9240-4484-96d0-8b5eb9e44ead"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.765398 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.765445 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.765459 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8be96a69-9240-4484-96d0-8b5eb9e44ead-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.776374 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-8sqvg"] Dec 08 15:05:08 crc kubenswrapper[4894]: I1208 15:05:08.787047 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-8sqvg"] Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.208440 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8be96a69-9240-4484-96d0-8b5eb9e44ead" path="/var/lib/kubelet/pods/8be96a69-9240-4484-96d0-8b5eb9e44ead/volumes" Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.253084 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5576bf896f-fxztt" Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.358034 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6889cd597d-q9lvm"] Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.358754 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6889cd597d-q9lvm" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-api" containerID="cri-o://c43962dbee949786a1ae21324ab487aef33edfb2e0c1fa13add22deff95cda81" gracePeriod=30 Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.360087 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6889cd597d-q9lvm" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-httpd" containerID="cri-o://02f2916c7384d69c38cfc90fa122b3d57d5b6a09dc89d3135906f2aa9067c004" gracePeriod=30 Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.432502 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="cinder-scheduler" containerID="cri-o://8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535" gracePeriod=30 Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.432969 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="probe" containerID="cri-o://402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d" gracePeriod=30 Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.519138 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:05:09 crc kubenswrapper[4894]: I1208 15:05:09.887773 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:05:10 crc kubenswrapper[4894]: I1208 15:05:10.172808 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7fb57bf44-rk2jn" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 08 15:05:10 crc kubenswrapper[4894]: I1208 15:05:10.452174 4894 generic.go:334] "Generic (PLEG): container finished" podID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerID="02f2916c7384d69c38cfc90fa122b3d57d5b6a09dc89d3135906f2aa9067c004" exitCode=0 Dec 08 15:05:10 crc kubenswrapper[4894]: I1208 15:05:10.452366 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6889cd597d-q9lvm" event={"ID":"4bba85bc-cd5a-4b17-8dff-339cb0f745a7","Type":"ContainerDied","Data":"02f2916c7384d69c38cfc90fa122b3d57d5b6a09dc89d3135906f2aa9067c004"} Dec 08 15:05:10 crc kubenswrapper[4894]: I1208 15:05:10.460843 4894 generic.go:334] "Generic (PLEG): container finished" podID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerID="31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844" exitCode=0 Dec 08 15:05:10 crc kubenswrapper[4894]: I1208 15:05:10.460975 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb57bf44-rk2jn" event={"ID":"a44fcb75-8381-44cf-8c5b-2b1108232807","Type":"ContainerDied","Data":"31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844"} Dec 08 15:05:10 crc kubenswrapper[4894]: I1208 15:05:10.568687 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:05:10 crc kubenswrapper[4894]: I1208 15:05:10.572863 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f74bf4b48-xdrm8" Dec 08 15:05:11 crc kubenswrapper[4894]: I1208 15:05:11.475656 4894 generic.go:334] "Generic (PLEG): container finished" podID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerID="402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d" exitCode=0 Dec 08 15:05:11 crc kubenswrapper[4894]: I1208 15:05:11.476544 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee3e5d72-93bc-437f-95eb-59ced7cb4590","Type":"ContainerDied","Data":"402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d"} Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.443042 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.538209 4894 generic.go:334] "Generic (PLEG): container finished" podID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerID="8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535" exitCode=0 Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.538313 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee3e5d72-93bc-437f-95eb-59ced7cb4590","Type":"ContainerDied","Data":"8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535"} Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.538347 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ee3e5d72-93bc-437f-95eb-59ced7cb4590","Type":"ContainerDied","Data":"739ac44faecfa962fdc5c5e632d561be7ea1839616d41655d390bc9ae924a543"} Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.538354 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.538368 4894 scope.go:117] "RemoveContainer" containerID="402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.567104 4894 generic.go:334] "Generic (PLEG): container finished" podID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerID="c43962dbee949786a1ae21324ab487aef33edfb2e0c1fa13add22deff95cda81" exitCode=0 Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.567174 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6889cd597d-q9lvm" event={"ID":"4bba85bc-cd5a-4b17-8dff-339cb0f745a7","Type":"ContainerDied","Data":"c43962dbee949786a1ae21324ab487aef33edfb2e0c1fa13add22deff95cda81"} Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.579371 4894 scope.go:117] "RemoveContainer" containerID="8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.604001 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-combined-ca-bundle\") pod \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.604143 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data\") pod \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.604214 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee3e5d72-93bc-437f-95eb-59ced7cb4590-etc-machine-id\") pod \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.604450 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data-custom\") pod \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.604495 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-scripts\") pod \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.604583 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghhhn\" (UniqueName: \"kubernetes.io/projected/ee3e5d72-93bc-437f-95eb-59ced7cb4590-kube-api-access-ghhhn\") pod \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\" (UID: \"ee3e5d72-93bc-437f-95eb-59ced7cb4590\") " Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.604940 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee3e5d72-93bc-437f-95eb-59ced7cb4590-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ee3e5d72-93bc-437f-95eb-59ced7cb4590" (UID: "ee3e5d72-93bc-437f-95eb-59ced7cb4590"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.606221 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee3e5d72-93bc-437f-95eb-59ced7cb4590-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.620025 4894 scope.go:117] "RemoveContainer" containerID="402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.620267 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee3e5d72-93bc-437f-95eb-59ced7cb4590-kube-api-access-ghhhn" (OuterVolumeSpecName: "kube-api-access-ghhhn") pod "ee3e5d72-93bc-437f-95eb-59ced7cb4590" (UID: "ee3e5d72-93bc-437f-95eb-59ced7cb4590"). InnerVolumeSpecName "kube-api-access-ghhhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.628068 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ee3e5d72-93bc-437f-95eb-59ced7cb4590" (UID: "ee3e5d72-93bc-437f-95eb-59ced7cb4590"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:12 crc kubenswrapper[4894]: E1208 15:05:12.631983 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d\": container with ID starting with 402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d not found: ID does not exist" containerID="402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.632029 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d"} err="failed to get container status \"402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d\": rpc error: code = NotFound desc = could not find container \"402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d\": container with ID starting with 402ff97e315a610180e3de1b598268429ca478a68fa056128e26aa1e6698566d not found: ID does not exist" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.632056 4894 scope.go:117] "RemoveContainer" containerID="8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535" Dec 08 15:05:12 crc kubenswrapper[4894]: E1208 15:05:12.636972 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535\": container with ID starting with 8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535 not found: ID does not exist" containerID="8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.637017 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535"} err="failed to get container status \"8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535\": rpc error: code = NotFound desc = could not find container \"8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535\": container with ID starting with 8576bbc98623b90149864a7a20f6509a3715c07c43cc0973c6dc8396d9a3b535 not found: ID does not exist" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.657220 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-scripts" (OuterVolumeSpecName: "scripts") pod "ee3e5d72-93bc-437f-95eb-59ced7cb4590" (UID: "ee3e5d72-93bc-437f-95eb-59ced7cb4590"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.712738 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.712774 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.712785 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghhhn\" (UniqueName: \"kubernetes.io/projected/ee3e5d72-93bc-437f-95eb-59ced7cb4590-kube-api-access-ghhhn\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.728974 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee3e5d72-93bc-437f-95eb-59ced7cb4590" (UID: "ee3e5d72-93bc-437f-95eb-59ced7cb4590"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.815071 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.864043 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data" (OuterVolumeSpecName: "config-data") pod "ee3e5d72-93bc-437f-95eb-59ced7cb4590" (UID: "ee3e5d72-93bc-437f-95eb-59ced7cb4590"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.919322 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee3e5d72-93bc-437f-95eb-59ced7cb4590-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:12 crc kubenswrapper[4894]: I1208 15:05:12.962325 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.121939 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqgp7\" (UniqueName: \"kubernetes.io/projected/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-kube-api-access-rqgp7\") pod \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.122020 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-ovndb-tls-certs\") pod \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.122104 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-config\") pod \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.122188 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-combined-ca-bundle\") pod \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.122264 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-httpd-config\") pod \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\" (UID: \"4bba85bc-cd5a-4b17-8dff-339cb0f745a7\") " Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.127584 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-kube-api-access-rqgp7" (OuterVolumeSpecName: "kube-api-access-rqgp7") pod "4bba85bc-cd5a-4b17-8dff-339cb0f745a7" (UID: "4bba85bc-cd5a-4b17-8dff-339cb0f745a7"). InnerVolumeSpecName "kube-api-access-rqgp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.143895 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "4bba85bc-cd5a-4b17-8dff-339cb0f745a7" (UID: "4bba85bc-cd5a-4b17-8dff-339cb0f745a7"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.193392 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.226793 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqgp7\" (UniqueName: \"kubernetes.io/projected/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-kube-api-access-rqgp7\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.226859 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.237889 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.249399 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-config" (OuterVolumeSpecName: "config") pod "4bba85bc-cd5a-4b17-8dff-339cb0f745a7" (UID: "4bba85bc-cd5a-4b17-8dff-339cb0f745a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.252934 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bba85bc-cd5a-4b17-8dff-339cb0f745a7" (UID: "4bba85bc-cd5a-4b17-8dff-339cb0f745a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.259026 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "4bba85bc-cd5a-4b17-8dff-339cb0f745a7" (UID: "4bba85bc-cd5a-4b17-8dff-339cb0f745a7"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.267202 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:05:13 crc kubenswrapper[4894]: E1208 15:05:13.267750 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="cinder-scheduler" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.267770 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="cinder-scheduler" Dec 08 15:05:13 crc kubenswrapper[4894]: E1208 15:05:13.267792 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-httpd" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.267799 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-httpd" Dec 08 15:05:13 crc kubenswrapper[4894]: E1208 15:05:13.267835 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="probe" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.267843 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="probe" Dec 08 15:05:13 crc kubenswrapper[4894]: E1208 15:05:13.267854 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerName="init" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.267862 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerName="init" Dec 08 15:05:13 crc kubenswrapper[4894]: E1208 15:05:13.267887 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerName="dnsmasq-dns" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.267896 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerName="dnsmasq-dns" Dec 08 15:05:13 crc kubenswrapper[4894]: E1208 15:05:13.267918 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-api" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.267925 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-api" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.268146 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-api" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.268169 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="cinder-scheduler" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.268183 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be96a69-9240-4484-96d0-8b5eb9e44ead" containerName="dnsmasq-dns" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.268201 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" containerName="neutron-httpd" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.268212 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" containerName="probe" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.269551 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.273184 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.280460 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.350545 4894 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.350581 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.350591 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bba85bc-cd5a-4b17-8dff-339cb0f745a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.452907 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-config-data\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.452975 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.453068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.453140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct4ff\" (UniqueName: \"kubernetes.io/projected/f2435a2b-5129-4479-b888-d3874b125874-kube-api-access-ct4ff\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.453191 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-scripts\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.453341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2435a2b-5129-4479-b888-d3874b125874-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.555844 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2435a2b-5129-4479-b888-d3874b125874-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.555959 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-config-data\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.555981 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.556018 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.556033 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2435a2b-5129-4479-b888-d3874b125874-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.556077 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct4ff\" (UniqueName: \"kubernetes.io/projected/f2435a2b-5129-4479-b888-d3874b125874-kube-api-access-ct4ff\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.556222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-scripts\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.560841 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-scripts\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.561656 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.562572 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-config-data\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.563096 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2435a2b-5129-4479-b888-d3874b125874-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.576570 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct4ff\" (UniqueName: \"kubernetes.io/projected/f2435a2b-5129-4479-b888-d3874b125874-kube-api-access-ct4ff\") pod \"cinder-scheduler-0\" (UID: \"f2435a2b-5129-4479-b888-d3874b125874\") " pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.584430 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6889cd597d-q9lvm" event={"ID":"4bba85bc-cd5a-4b17-8dff-339cb0f745a7","Type":"ContainerDied","Data":"780ffb924992ee6d559234abe082b4a97406b0071a3d409c81d514b7e28eef32"} Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.584503 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6889cd597d-q9lvm" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.584518 4894 scope.go:117] "RemoveContainer" containerID="02f2916c7384d69c38cfc90fa122b3d57d5b6a09dc89d3135906f2aa9067c004" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.609651 4894 scope.go:117] "RemoveContainer" containerID="c43962dbee949786a1ae21324ab487aef33edfb2e0c1fa13add22deff95cda81" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.609804 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.654714 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6889cd597d-q9lvm"] Dec 08 15:05:13 crc kubenswrapper[4894]: I1208 15:05:13.681479 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6889cd597d-q9lvm"] Dec 08 15:05:14 crc kubenswrapper[4894]: I1208 15:05:14.253971 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 08 15:05:14 crc kubenswrapper[4894]: W1208 15:05:14.274361 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2435a2b_5129_4479_b888_d3874b125874.slice/crio-504400b508690c67e9b0a877b1688e6419f3baa264535420a0bed5b97f96ad71 WatchSource:0}: Error finding container 504400b508690c67e9b0a877b1688e6419f3baa264535420a0bed5b97f96ad71: Status 404 returned error can't find the container with id 504400b508690c67e9b0a877b1688e6419f3baa264535420a0bed5b97f96ad71 Dec 08 15:05:14 crc kubenswrapper[4894]: I1208 15:05:14.352521 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7cb79b8b98-42mc8" Dec 08 15:05:14 crc kubenswrapper[4894]: I1208 15:05:14.610485 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2435a2b-5129-4479-b888-d3874b125874","Type":"ContainerStarted","Data":"504400b508690c67e9b0a877b1688e6419f3baa264535420a0bed5b97f96ad71"} Dec 08 15:05:15 crc kubenswrapper[4894]: I1208 15:05:15.210712 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bba85bc-cd5a-4b17-8dff-339cb0f745a7" path="/var/lib/kubelet/pods/4bba85bc-cd5a-4b17-8dff-339cb0f745a7/volumes" Dec 08 15:05:15 crc kubenswrapper[4894]: I1208 15:05:15.212002 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee3e5d72-93bc-437f-95eb-59ced7cb4590" path="/var/lib/kubelet/pods/ee3e5d72-93bc-437f-95eb-59ced7cb4590/volumes" Dec 08 15:05:15 crc kubenswrapper[4894]: I1208 15:05:15.661030 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2435a2b-5129-4479-b888-d3874b125874","Type":"ContainerStarted","Data":"36c689f0d2eafc5ca9df2c3956cc0161d1fcd283e3078e4cb6110a7fb5df0a11"} Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.382971 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.384305 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.386794 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.387114 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.387133 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zc6rg" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.395122 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.529568 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9fpk\" (UniqueName: \"kubernetes.io/projected/29e0f64a-aac1-4909-b7f6-d83223c04d31-kube-api-access-d9fpk\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.530001 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29e0f64a-aac1-4909-b7f6-d83223c04d31-openstack-config-secret\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.530070 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29e0f64a-aac1-4909-b7f6-d83223c04d31-openstack-config\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.530120 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e0f64a-aac1-4909-b7f6-d83223c04d31-combined-ca-bundle\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.631673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9fpk\" (UniqueName: \"kubernetes.io/projected/29e0f64a-aac1-4909-b7f6-d83223c04d31-kube-api-access-d9fpk\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.631757 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29e0f64a-aac1-4909-b7f6-d83223c04d31-openstack-config-secret\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.631809 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29e0f64a-aac1-4909-b7f6-d83223c04d31-openstack-config\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.631874 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e0f64a-aac1-4909-b7f6-d83223c04d31-combined-ca-bundle\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.633743 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/29e0f64a-aac1-4909-b7f6-d83223c04d31-openstack-config\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.641080 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/29e0f64a-aac1-4909-b7f6-d83223c04d31-openstack-config-secret\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.648694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e0f64a-aac1-4909-b7f6-d83223c04d31-combined-ca-bundle\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.656710 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9fpk\" (UniqueName: \"kubernetes.io/projected/29e0f64a-aac1-4909-b7f6-d83223c04d31-kube-api-access-d9fpk\") pod \"openstackclient\" (UID: \"29e0f64a-aac1-4909-b7f6-d83223c04d31\") " pod="openstack/openstackclient" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.680576 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f2435a2b-5129-4479-b888-d3874b125874","Type":"ContainerStarted","Data":"5783bdac65cffdd0989bb238d65fe3bf467fa85aa5f429fc609ac47e473750a8"} Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.714471 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.714433879 podStartE2EDuration="3.714433879s" podCreationTimestamp="2025-12-08 15:05:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:16.709978899 +0000 UTC m=+1137.809985024" watchObservedRunningTime="2025-12-08 15:05:16.714433879 +0000 UTC m=+1137.814440004" Dec 08 15:05:16 crc kubenswrapper[4894]: I1208 15:05:16.718235 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 08 15:05:17 crc kubenswrapper[4894]: I1208 15:05:17.002523 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:17 crc kubenswrapper[4894]: I1208 15:05:17.456248 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 08 15:05:17 crc kubenswrapper[4894]: W1208 15:05:17.468050 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29e0f64a_aac1_4909_b7f6_d83223c04d31.slice/crio-f714ed272b93104530e7dcddc1a8e4b43893bb290115347c06264504becf57cd WatchSource:0}: Error finding container f714ed272b93104530e7dcddc1a8e4b43893bb290115347c06264504becf57cd: Status 404 returned error can't find the container with id f714ed272b93104530e7dcddc1a8e4b43893bb290115347c06264504becf57cd Dec 08 15:05:17 crc kubenswrapper[4894]: I1208 15:05:17.729110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"29e0f64a-aac1-4909-b7f6-d83223c04d31","Type":"ContainerStarted","Data":"f714ed272b93104530e7dcddc1a8e4b43893bb290115347c06264504becf57cd"} Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.113412 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.171757 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-789cf67ffd-s9792" Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.248039 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-86796dcf68-kb8nc"] Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.248285 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-86796dcf68-kb8nc" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api-log" containerID="cri-o://f1d1c24cc60914a0213b76ff2b12244b7908f3edb4f8aae3c27c5b1365c9bb94" gracePeriod=30 Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.248486 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-86796dcf68-kb8nc" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api" containerID="cri-o://95e118df7c94651c2914923cdf75d03fa41f501e562cabef577a60f94a0ee1ab" gracePeriod=30 Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.611245 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.755029 4894 generic.go:334] "Generic (PLEG): container finished" podID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerID="f1d1c24cc60914a0213b76ff2b12244b7908f3edb4f8aae3c27c5b1365c9bb94" exitCode=143 Dec 08 15:05:18 crc kubenswrapper[4894]: I1208 15:05:18.756392 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86796dcf68-kb8nc" event={"ID":"1805e51a-2b46-4c58-890c-8bfe560b74f3","Type":"ContainerDied","Data":"f1d1c24cc60914a0213b76ff2b12244b7908f3edb4f8aae3c27c5b1365c9bb94"} Dec 08 15:05:20 crc kubenswrapper[4894]: I1208 15:05:20.169914 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7fb57bf44-rk2jn" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 08 15:05:21 crc kubenswrapper[4894]: I1208 15:05:21.788678 4894 generic.go:334] "Generic (PLEG): container finished" podID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerID="95e118df7c94651c2914923cdf75d03fa41f501e562cabef577a60f94a0ee1ab" exitCode=0 Dec 08 15:05:21 crc kubenswrapper[4894]: I1208 15:05:21.788755 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86796dcf68-kb8nc" event={"ID":"1805e51a-2b46-4c58-890c-8bfe560b74f3","Type":"ContainerDied","Data":"95e118df7c94651c2914923cdf75d03fa41f501e562cabef577a60f94a0ee1ab"} Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.381586 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.481270 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4rxl\" (UniqueName: \"kubernetes.io/projected/1805e51a-2b46-4c58-890c-8bfe560b74f3-kube-api-access-z4rxl\") pod \"1805e51a-2b46-4c58-890c-8bfe560b74f3\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.481315 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data-custom\") pod \"1805e51a-2b46-4c58-890c-8bfe560b74f3\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.481348 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-combined-ca-bundle\") pod \"1805e51a-2b46-4c58-890c-8bfe560b74f3\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.481491 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data\") pod \"1805e51a-2b46-4c58-890c-8bfe560b74f3\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.481595 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1805e51a-2b46-4c58-890c-8bfe560b74f3-logs\") pod \"1805e51a-2b46-4c58-890c-8bfe560b74f3\" (UID: \"1805e51a-2b46-4c58-890c-8bfe560b74f3\") " Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.482493 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1805e51a-2b46-4c58-890c-8bfe560b74f3-logs" (OuterVolumeSpecName: "logs") pod "1805e51a-2b46-4c58-890c-8bfe560b74f3" (UID: "1805e51a-2b46-4c58-890c-8bfe560b74f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.486369 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1805e51a-2b46-4c58-890c-8bfe560b74f3-kube-api-access-z4rxl" (OuterVolumeSpecName: "kube-api-access-z4rxl") pod "1805e51a-2b46-4c58-890c-8bfe560b74f3" (UID: "1805e51a-2b46-4c58-890c-8bfe560b74f3"). InnerVolumeSpecName "kube-api-access-z4rxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.486577 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1805e51a-2b46-4c58-890c-8bfe560b74f3" (UID: "1805e51a-2b46-4c58-890c-8bfe560b74f3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.514601 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1805e51a-2b46-4c58-890c-8bfe560b74f3" (UID: "1805e51a-2b46-4c58-890c-8bfe560b74f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.547076 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data" (OuterVolumeSpecName: "config-data") pod "1805e51a-2b46-4c58-890c-8bfe560b74f3" (UID: "1805e51a-2b46-4c58-890c-8bfe560b74f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.583971 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4rxl\" (UniqueName: \"kubernetes.io/projected/1805e51a-2b46-4c58-890c-8bfe560b74f3-kube-api-access-z4rxl\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.584002 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.584012 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.584021 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1805e51a-2b46-4c58-890c-8bfe560b74f3-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.584031 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1805e51a-2b46-4c58-890c-8bfe560b74f3-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.800718 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-86796dcf68-kb8nc" event={"ID":"1805e51a-2b46-4c58-890c-8bfe560b74f3","Type":"ContainerDied","Data":"19144d8a407ae983e1b1df1ee98c6c03634257a15a97bf0c44cee1817aee3479"} Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.800783 4894 scope.go:117] "RemoveContainer" containerID="95e118df7c94651c2914923cdf75d03fa41f501e562cabef577a60f94a0ee1ab" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.801625 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-86796dcf68-kb8nc" Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.839900 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-86796dcf68-kb8nc"] Dec 08 15:05:22 crc kubenswrapper[4894]: I1208 15:05:22.851538 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-86796dcf68-kb8nc"] Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.210259 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" path="/var/lib/kubelet/pods/1805e51a-2b46-4c58-890c-8bfe560b74f3/volumes" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.350021 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-567d6f4bf7-xrnjd"] Dec 08 15:05:23 crc kubenswrapper[4894]: E1208 15:05:23.350399 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.350415 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api" Dec 08 15:05:23 crc kubenswrapper[4894]: E1208 15:05:23.350427 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api-log" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.350434 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api-log" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.350648 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.350673 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1805e51a-2b46-4c58-890c-8bfe560b74f3" containerName="barbican-api-log" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.351607 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.356439 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.356492 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.357684 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.367775 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-567d6f4bf7-xrnjd"] Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.500791 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-combined-ca-bundle\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.501077 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-log-httpd\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.501131 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-run-httpd\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.501221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-config-data\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.501254 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-public-tls-certs\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.501288 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-etc-swift\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.501309 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-internal-tls-certs\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.501327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm9g9\" (UniqueName: \"kubernetes.io/projected/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-kube-api-access-sm9g9\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.602875 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-log-httpd\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.602930 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-run-httpd\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.602977 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-config-data\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.603000 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-public-tls-certs\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.603022 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-etc-swift\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.603106 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-internal-tls-certs\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.603129 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm9g9\" (UniqueName: \"kubernetes.io/projected/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-kube-api-access-sm9g9\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.603166 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-combined-ca-bundle\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.603536 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-run-httpd\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.603993 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-log-httpd\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.610112 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-internal-tls-certs\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.610467 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-etc-swift\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.613443 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-public-tls-certs\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.613684 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-combined-ca-bundle\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.614041 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-config-data\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.628652 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm9g9\" (UniqueName: \"kubernetes.io/projected/c7a0f531-0d8b-4858-aee9-981bdfbc81e0-kube-api-access-sm9g9\") pod \"swift-proxy-567d6f4bf7-xrnjd\" (UID: \"c7a0f531-0d8b-4858-aee9-981bdfbc81e0\") " pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.670347 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:23 crc kubenswrapper[4894]: I1208 15:05:23.932727 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.602708 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.603236 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-central-agent" containerID="cri-o://9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369" gracePeriod=30 Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.603416 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="proxy-httpd" containerID="cri-o://2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f" gracePeriod=30 Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.603534 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-notification-agent" containerID="cri-o://61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b" gracePeriod=30 Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.603604 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="sg-core" containerID="cri-o://7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21" gracePeriod=30 Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.624756 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.163:3000/\": EOF" Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.823747 4894 generic.go:334] "Generic (PLEG): container finished" podID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerID="7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21" exitCode=2 Dec 08 15:05:24 crc kubenswrapper[4894]: I1208 15:05:24.823801 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerDied","Data":"7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21"} Dec 08 15:05:25 crc kubenswrapper[4894]: I1208 15:05:25.872646 4894 generic.go:334] "Generic (PLEG): container finished" podID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerID="2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f" exitCode=0 Dec 08 15:05:25 crc kubenswrapper[4894]: I1208 15:05:25.872678 4894 generic.go:334] "Generic (PLEG): container finished" podID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerID="9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369" exitCode=0 Dec 08 15:05:25 crc kubenswrapper[4894]: I1208 15:05:25.872697 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerDied","Data":"2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f"} Dec 08 15:05:25 crc kubenswrapper[4894]: I1208 15:05:25.872721 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerDied","Data":"9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369"} Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.475609 4894 scope.go:117] "RemoveContainer" containerID="f1d1c24cc60914a0213b76ff2b12244b7908f3edb4f8aae3c27c5b1365c9bb94" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.834584 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.908160 4894 generic.go:334] "Generic (PLEG): container finished" podID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerID="61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b" exitCode=0 Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.908254 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerDied","Data":"61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b"} Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.908282 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f25b4581-989b-4f47-b680-063ea7eb57f7","Type":"ContainerDied","Data":"fe7e26deba89966a0467f3ad00ea49c4ed6c3b0cf2e828a2a2a5916feda1213a"} Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.908300 4894 scope.go:117] "RemoveContainer" containerID="2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.908430 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.909531 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-scripts\") pod \"f25b4581-989b-4f47-b680-063ea7eb57f7\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.909563 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-run-httpd\") pod \"f25b4581-989b-4f47-b680-063ea7eb57f7\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.909619 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-log-httpd\") pod \"f25b4581-989b-4f47-b680-063ea7eb57f7\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.909749 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-config-data\") pod \"f25b4581-989b-4f47-b680-063ea7eb57f7\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.909771 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-combined-ca-bundle\") pod \"f25b4581-989b-4f47-b680-063ea7eb57f7\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.909794 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x74d\" (UniqueName: \"kubernetes.io/projected/f25b4581-989b-4f47-b680-063ea7eb57f7-kube-api-access-6x74d\") pod \"f25b4581-989b-4f47-b680-063ea7eb57f7\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.909856 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-sg-core-conf-yaml\") pod \"f25b4581-989b-4f47-b680-063ea7eb57f7\" (UID: \"f25b4581-989b-4f47-b680-063ea7eb57f7\") " Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.910327 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f25b4581-989b-4f47-b680-063ea7eb57f7" (UID: "f25b4581-989b-4f47-b680-063ea7eb57f7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.911713 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f25b4581-989b-4f47-b680-063ea7eb57f7" (UID: "f25b4581-989b-4f47-b680-063ea7eb57f7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.916416 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"29e0f64a-aac1-4909-b7f6-d83223c04d31","Type":"ContainerStarted","Data":"dbffd2b805f2ca590bd44c4ca11f2a75c38fec2784ddc123caa7e4d99db6592a"} Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.916975 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f25b4581-989b-4f47-b680-063ea7eb57f7-kube-api-access-6x74d" (OuterVolumeSpecName: "kube-api-access-6x74d") pod "f25b4581-989b-4f47-b680-063ea7eb57f7" (UID: "f25b4581-989b-4f47-b680-063ea7eb57f7"). InnerVolumeSpecName "kube-api-access-6x74d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.925983 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-scripts" (OuterVolumeSpecName: "scripts") pod "f25b4581-989b-4f47-b680-063ea7eb57f7" (UID: "f25b4581-989b-4f47-b680-063ea7eb57f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.945831 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.846074241 podStartE2EDuration="12.94577315s" podCreationTimestamp="2025-12-08 15:05:16 +0000 UTC" firstStartedPulling="2025-12-08 15:05:17.470464287 +0000 UTC m=+1138.570470402" lastFinishedPulling="2025-12-08 15:05:28.570163196 +0000 UTC m=+1149.670169311" observedRunningTime="2025-12-08 15:05:28.939589706 +0000 UTC m=+1150.039595841" watchObservedRunningTime="2025-12-08 15:05:28.94577315 +0000 UTC m=+1150.045779275" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.947007 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f25b4581-989b-4f47-b680-063ea7eb57f7" (UID: "f25b4581-989b-4f47-b680-063ea7eb57f7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.952261 4894 scope.go:117] "RemoveContainer" containerID="7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21" Dec 08 15:05:28 crc kubenswrapper[4894]: I1208 15:05:28.979575 4894 scope.go:117] "RemoveContainer" containerID="61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.013253 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.013305 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x74d\" (UniqueName: \"kubernetes.io/projected/f25b4581-989b-4f47-b680-063ea7eb57f7-kube-api-access-6x74d\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.013320 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.013332 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.013341 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f25b4581-989b-4f47-b680-063ea7eb57f7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.019471 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f25b4581-989b-4f47-b680-063ea7eb57f7" (UID: "f25b4581-989b-4f47-b680-063ea7eb57f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.020553 4894 scope.go:117] "RemoveContainer" containerID="9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.048046 4894 scope.go:117] "RemoveContainer" containerID="2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f" Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.048605 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f\": container with ID starting with 2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f not found: ID does not exist" containerID="2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.048669 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f"} err="failed to get container status \"2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f\": rpc error: code = NotFound desc = could not find container \"2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f\": container with ID starting with 2d471d370aea2fb9fc398cfee23a53318fa7452b0aea8a913bed46cb3d22628f not found: ID does not exist" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.048716 4894 scope.go:117] "RemoveContainer" containerID="7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21" Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.049427 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21\": container with ID starting with 7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21 not found: ID does not exist" containerID="7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.049457 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21"} err="failed to get container status \"7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21\": rpc error: code = NotFound desc = could not find container \"7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21\": container with ID starting with 7c66317fccce91994d203b615d1b176a3187bb40df5bf9255ba5db93415b9d21 not found: ID does not exist" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.049473 4894 scope.go:117] "RemoveContainer" containerID="61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b" Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.050034 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b\": container with ID starting with 61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b not found: ID does not exist" containerID="61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.050065 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b"} err="failed to get container status \"61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b\": rpc error: code = NotFound desc = could not find container \"61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b\": container with ID starting with 61d439b9d29e44d9d8a4dd96be8821cf7074e1b0625402732f6dfff0b303fd3b not found: ID does not exist" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.050084 4894 scope.go:117] "RemoveContainer" containerID="9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369" Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.050585 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369\": container with ID starting with 9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369 not found: ID does not exist" containerID="9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.050630 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369"} err="failed to get container status \"9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369\": rpc error: code = NotFound desc = could not find container \"9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369\": container with ID starting with 9f0657137ff4c09dbff59d5e9c2339d30097f0a6f740b38843e6546a87293369 not found: ID does not exist" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.060755 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-config-data" (OuterVolumeSpecName: "config-data") pod "f25b4581-989b-4f47-b680-063ea7eb57f7" (UID: "f25b4581-989b-4f47-b680-063ea7eb57f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.115540 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.115583 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f25b4581-989b-4f47-b680-063ea7eb57f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.184043 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-567d6f4bf7-xrnjd"] Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.315444 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.325302 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.334389 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.335052 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-notification-agent" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335072 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-notification-agent" Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.335091 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="proxy-httpd" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335097 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="proxy-httpd" Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.335128 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="sg-core" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335133 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="sg-core" Dec 08 15:05:29 crc kubenswrapper[4894]: E1208 15:05:29.335143 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-central-agent" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335149 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-central-agent" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335410 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-notification-agent" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335435 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="sg-core" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335448 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="ceilometer-central-agent" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.335461 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" containerName="proxy-httpd" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.337854 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.341446 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.341666 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.347456 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.422848 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-config-data\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.422961 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.422984 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.423030 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-log-httpd\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.423089 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-run-httpd\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.423124 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-scripts\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.423151 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l56xs\" (UniqueName: \"kubernetes.io/projected/9c952e58-c5cb-494c-89fe-b34c835ba0d5-kube-api-access-l56xs\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.526767 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-run-httpd\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.527420 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-run-httpd\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.527435 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-scripts\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.527525 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l56xs\" (UniqueName: \"kubernetes.io/projected/9c952e58-c5cb-494c-89fe-b34c835ba0d5-kube-api-access-l56xs\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.527684 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-config-data\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.527897 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.527923 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.528057 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-log-httpd\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.528476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-log-httpd\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.535089 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-scripts\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.535234 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-config-data\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.538846 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.541365 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.552177 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l56xs\" (UniqueName: \"kubernetes.io/projected/9c952e58-c5cb-494c-89fe-b34c835ba0d5-kube-api-access-l56xs\") pod \"ceilometer-0\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.676546 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.944165 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" event={"ID":"c7a0f531-0d8b-4858-aee9-981bdfbc81e0","Type":"ContainerStarted","Data":"60f204749e1a53ef09aae6b8a0864d89cbda01c0e4845ca31a33719a12c5f307"} Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.945698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" event={"ID":"c7a0f531-0d8b-4858-aee9-981bdfbc81e0","Type":"ContainerStarted","Data":"89f51c339798480b2690cd64d4dcf6f93c473efd3d1daf823f9534f7827ebe8d"} Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.945731 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" event={"ID":"c7a0f531-0d8b-4858-aee9-981bdfbc81e0","Type":"ContainerStarted","Data":"c73060c1527f3f6c172bca4be0b0abd34c34db3bb1067f422b38733860b6b2aa"} Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.945754 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:29 crc kubenswrapper[4894]: I1208 15:05:29.945769 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:30 crc kubenswrapper[4894]: I1208 15:05:30.169720 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7fb57bf44-rk2jn" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 08 15:05:30 crc kubenswrapper[4894]: I1208 15:05:30.169926 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:05:30 crc kubenswrapper[4894]: I1208 15:05:30.190549 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" podStartSLOduration=7.190523824 podStartE2EDuration="7.190523824s" podCreationTimestamp="2025-12-08 15:05:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:29.972156545 +0000 UTC m=+1151.072162660" watchObservedRunningTime="2025-12-08 15:05:30.190523824 +0000 UTC m=+1151.290529939" Dec 08 15:05:30 crc kubenswrapper[4894]: I1208 15:05:30.202578 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:30 crc kubenswrapper[4894]: I1208 15:05:30.406934 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:30 crc kubenswrapper[4894]: I1208 15:05:30.953776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerStarted","Data":"e1fcad5445dedb315a13b1bff2cd224d12e543c46865df09f708d936e1fe7444"} Dec 08 15:05:31 crc kubenswrapper[4894]: I1208 15:05:31.207720 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f25b4581-989b-4f47-b680-063ea7eb57f7" path="/var/lib/kubelet/pods/f25b4581-989b-4f47-b680-063ea7eb57f7/volumes" Dec 08 15:05:31 crc kubenswrapper[4894]: I1208 15:05:31.966691 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerStarted","Data":"f8c8b14267e697c651c7fa31eb256e5d6f570520e0c6699cba6d0ed0a78d748b"} Dec 08 15:05:31 crc kubenswrapper[4894]: I1208 15:05:31.966773 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerStarted","Data":"b01a17b6468f37c6509b24f9a48eb8bc7d7115b0182768368ea8b4dd06ecaafb"} Dec 08 15:05:32 crc kubenswrapper[4894]: I1208 15:05:32.983644 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerStarted","Data":"1d51d7ecc73ef5a36a2ec098b973f0156cb9ccb60573b6daba9745b8a00d2ba8"} Dec 08 15:05:35 crc kubenswrapper[4894]: I1208 15:05:35.002825 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerStarted","Data":"fd084a1e25b30a9c177155143cb7f567f62a386774b4089ed7f34cdf3e83fd13"} Dec 08 15:05:35 crc kubenswrapper[4894]: I1208 15:05:35.002992 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-central-agent" containerID="cri-o://b01a17b6468f37c6509b24f9a48eb8bc7d7115b0182768368ea8b4dd06ecaafb" gracePeriod=30 Dec 08 15:05:35 crc kubenswrapper[4894]: I1208 15:05:35.003043 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="sg-core" containerID="cri-o://1d51d7ecc73ef5a36a2ec098b973f0156cb9ccb60573b6daba9745b8a00d2ba8" gracePeriod=30 Dec 08 15:05:35 crc kubenswrapper[4894]: I1208 15:05:35.003022 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="proxy-httpd" containerID="cri-o://fd084a1e25b30a9c177155143cb7f567f62a386774b4089ed7f34cdf3e83fd13" gracePeriod=30 Dec 08 15:05:35 crc kubenswrapper[4894]: I1208 15:05:35.003065 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-notification-agent" containerID="cri-o://f8c8b14267e697c651c7fa31eb256e5d6f570520e0c6699cba6d0ed0a78d748b" gracePeriod=30 Dec 08 15:05:35 crc kubenswrapper[4894]: I1208 15:05:35.003223 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:05:35 crc kubenswrapper[4894]: I1208 15:05:35.030121 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.199276587 podStartE2EDuration="6.030104783s" podCreationTimestamp="2025-12-08 15:05:29 +0000 UTC" firstStartedPulling="2025-12-08 15:05:30.20283991 +0000 UTC m=+1151.302846015" lastFinishedPulling="2025-12-08 15:05:34.033668096 +0000 UTC m=+1155.133674211" observedRunningTime="2025-12-08 15:05:35.027869493 +0000 UTC m=+1156.127875628" watchObservedRunningTime="2025-12-08 15:05:35.030104783 +0000 UTC m=+1156.130110898" Dec 08 15:05:36 crc kubenswrapper[4894]: I1208 15:05:36.014251 4894 generic.go:334] "Generic (PLEG): container finished" podID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerID="fd084a1e25b30a9c177155143cb7f567f62a386774b4089ed7f34cdf3e83fd13" exitCode=0 Dec 08 15:05:36 crc kubenswrapper[4894]: I1208 15:05:36.014513 4894 generic.go:334] "Generic (PLEG): container finished" podID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerID="1d51d7ecc73ef5a36a2ec098b973f0156cb9ccb60573b6daba9745b8a00d2ba8" exitCode=2 Dec 08 15:05:36 crc kubenswrapper[4894]: I1208 15:05:36.014522 4894 generic.go:334] "Generic (PLEG): container finished" podID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerID="f8c8b14267e697c651c7fa31eb256e5d6f570520e0c6699cba6d0ed0a78d748b" exitCode=0 Dec 08 15:05:36 crc kubenswrapper[4894]: I1208 15:05:36.014548 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerDied","Data":"fd084a1e25b30a9c177155143cb7f567f62a386774b4089ed7f34cdf3e83fd13"} Dec 08 15:05:36 crc kubenswrapper[4894]: I1208 15:05:36.014575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerDied","Data":"1d51d7ecc73ef5a36a2ec098b973f0156cb9ccb60573b6daba9745b8a00d2ba8"} Dec 08 15:05:36 crc kubenswrapper[4894]: I1208 15:05:36.014584 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerDied","Data":"f8c8b14267e697c651c7fa31eb256e5d6f570520e0c6699cba6d0ed0a78d748b"} Dec 08 15:05:36 crc kubenswrapper[4894]: I1208 15:05:36.887391 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.006727 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a44fcb75-8381-44cf-8c5b-2b1108232807-logs\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.007387 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a44fcb75-8381-44cf-8c5b-2b1108232807-logs" (OuterVolumeSpecName: "logs") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.007454 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-combined-ca-bundle\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.007681 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzd27\" (UniqueName: \"kubernetes.io/projected/a44fcb75-8381-44cf-8c5b-2b1108232807-kube-api-access-qzd27\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.007787 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-config-data\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.007810 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-tls-certs\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.007863 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.008018 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-secret-key\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.008613 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a44fcb75-8381-44cf-8c5b-2b1108232807-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.058809 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a44fcb75-8381-44cf-8c5b-2b1108232807-kube-api-access-qzd27" (OuterVolumeSpecName: "kube-api-access-qzd27") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "kube-api-access-qzd27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.059136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.078966 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-config-data" (OuterVolumeSpecName: "config-data") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.097035 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.122428 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.122449 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.122459 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.122467 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzd27\" (UniqueName: \"kubernetes.io/projected/a44fcb75-8381-44cf-8c5b-2b1108232807-kube-api-access-qzd27\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.179114 4894 generic.go:334] "Generic (PLEG): container finished" podID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerID="b01a17b6468f37c6509b24f9a48eb8bc7d7115b0182768368ea8b4dd06ecaafb" exitCode=0 Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.179206 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerDied","Data":"b01a17b6468f37c6509b24f9a48eb8bc7d7115b0182768368ea8b4dd06ecaafb"} Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.229486 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts" (OuterVolumeSpecName: "scripts") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.232096 4894 generic.go:334] "Generic (PLEG): container finished" podID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerID="49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7" exitCode=137 Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.232200 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb57bf44-rk2jn" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.235120 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts\") pod \"a44fcb75-8381-44cf-8c5b-2b1108232807\" (UID: \"a44fcb75-8381-44cf-8c5b-2b1108232807\") " Dec 08 15:05:37 crc kubenswrapper[4894]: W1208 15:05:37.235887 4894 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a44fcb75-8381-44cf-8c5b-2b1108232807/volumes/kubernetes.io~configmap/scripts Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.235904 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts" (OuterVolumeSpecName: "scripts") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.261700 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb57bf44-rk2jn" event={"ID":"a44fcb75-8381-44cf-8c5b-2b1108232807","Type":"ContainerDied","Data":"49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7"} Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.261747 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb57bf44-rk2jn" event={"ID":"a44fcb75-8381-44cf-8c5b-2b1108232807","Type":"ContainerDied","Data":"be906960d703b7c0096a82357a5f2d9e67c1c8a1579ff7d541bb3e01f93bc5ff"} Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.261785 4894 scope.go:117] "RemoveContainer" containerID="31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.279291 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "a44fcb75-8381-44cf-8c5b-2b1108232807" (UID: "a44fcb75-8381-44cf-8c5b-2b1108232807"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.297979 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.298046 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.303398 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.341688 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44fcb75-8381-44cf-8c5b-2b1108232807-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.341746 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a44fcb75-8381-44cf-8c5b-2b1108232807-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.443860 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-log-httpd\") pod \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.443958 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-scripts\") pod \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.443994 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-run-httpd\") pod \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.444153 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-config-data\") pod \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.444180 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l56xs\" (UniqueName: \"kubernetes.io/projected/9c952e58-c5cb-494c-89fe-b34c835ba0d5-kube-api-access-l56xs\") pod \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.444196 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-combined-ca-bundle\") pod \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.444274 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-sg-core-conf-yaml\") pod \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\" (UID: \"9c952e58-c5cb-494c-89fe-b34c835ba0d5\") " Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.444742 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9c952e58-c5cb-494c-89fe-b34c835ba0d5" (UID: "9c952e58-c5cb-494c-89fe-b34c835ba0d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.451151 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-scripts" (OuterVolumeSpecName: "scripts") pod "9c952e58-c5cb-494c-89fe-b34c835ba0d5" (UID: "9c952e58-c5cb-494c-89fe-b34c835ba0d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.452100 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c952e58-c5cb-494c-89fe-b34c835ba0d5-kube-api-access-l56xs" (OuterVolumeSpecName: "kube-api-access-l56xs") pod "9c952e58-c5cb-494c-89fe-b34c835ba0d5" (UID: "9c952e58-c5cb-494c-89fe-b34c835ba0d5"). InnerVolumeSpecName "kube-api-access-l56xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.452449 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9c952e58-c5cb-494c-89fe-b34c835ba0d5" (UID: "9c952e58-c5cb-494c-89fe-b34c835ba0d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.480614 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9c952e58-c5cb-494c-89fe-b34c835ba0d5" (UID: "9c952e58-c5cb-494c-89fe-b34c835ba0d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.496782 4894 scope.go:117] "RemoveContainer" containerID="49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.540083 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c952e58-c5cb-494c-89fe-b34c835ba0d5" (UID: "9c952e58-c5cb-494c-89fe-b34c835ba0d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.547319 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l56xs\" (UniqueName: \"kubernetes.io/projected/9c952e58-c5cb-494c-89fe-b34c835ba0d5-kube-api-access-l56xs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.547539 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.547615 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.547730 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.547806 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.547912 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9c952e58-c5cb-494c-89fe-b34c835ba0d5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.569088 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-config-data" (OuterVolumeSpecName: "config-data") pod "9c952e58-c5cb-494c-89fe-b34c835ba0d5" (UID: "9c952e58-c5cb-494c-89fe-b34c835ba0d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.630056 4894 scope.go:117] "RemoveContainer" containerID="31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844" Dec 08 15:05:37 crc kubenswrapper[4894]: E1208 15:05:37.630603 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844\": container with ID starting with 31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844 not found: ID does not exist" containerID="31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.630915 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844"} err="failed to get container status \"31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844\": rpc error: code = NotFound desc = could not find container \"31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844\": container with ID starting with 31fd3a4f9d61527c8105ffa3f94d841dea76088188d53ac04f091aa6b6d9e844 not found: ID does not exist" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.630949 4894 scope.go:117] "RemoveContainer" containerID="49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7" Dec 08 15:05:37 crc kubenswrapper[4894]: E1208 15:05:37.631394 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7\": container with ID starting with 49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7 not found: ID does not exist" containerID="49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.631418 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7"} err="failed to get container status \"49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7\": rpc error: code = NotFound desc = could not find container \"49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7\": container with ID starting with 49e3053ffec08bda4bc962d972dc091985a8ce21a3ac430c12e91e091a4d48e7 not found: ID does not exist" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.644558 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fb57bf44-rk2jn"] Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.649182 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c952e58-c5cb-494c-89fe-b34c835ba0d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:37 crc kubenswrapper[4894]: I1208 15:05:37.652997 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7fb57bf44-rk2jn"] Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.245032 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9c952e58-c5cb-494c-89fe-b34c835ba0d5","Type":"ContainerDied","Data":"e1fcad5445dedb315a13b1bff2cd224d12e543c46865df09f708d936e1fe7444"} Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.245062 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.245104 4894 scope.go:117] "RemoveContainer" containerID="fd084a1e25b30a9c177155143cb7f567f62a386774b4089ed7f34cdf3e83fd13" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.263975 4894 scope.go:117] "RemoveContainer" containerID="1d51d7ecc73ef5a36a2ec098b973f0156cb9ccb60573b6daba9745b8a00d2ba8" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.290381 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.295064 4894 scope.go:117] "RemoveContainer" containerID="f8c8b14267e697c651c7fa31eb256e5d6f570520e0c6699cba6d0ed0a78d748b" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.303150 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321149 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321238 4894 scope.go:117] "RemoveContainer" containerID="b01a17b6468f37c6509b24f9a48eb8bc7d7115b0182768368ea8b4dd06ecaafb" Dec 08 15:05:38 crc kubenswrapper[4894]: E1208 15:05:38.321809 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="proxy-httpd" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321839 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="proxy-httpd" Dec 08 15:05:38 crc kubenswrapper[4894]: E1208 15:05:38.321865 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="sg-core" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321871 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="sg-core" Dec 08 15:05:38 crc kubenswrapper[4894]: E1208 15:05:38.321883 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-notification-agent" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321890 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-notification-agent" Dec 08 15:05:38 crc kubenswrapper[4894]: E1208 15:05:38.321905 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-central-agent" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321911 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-central-agent" Dec 08 15:05:38 crc kubenswrapper[4894]: E1208 15:05:38.321923 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321929 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" Dec 08 15:05:38 crc kubenswrapper[4894]: E1208 15:05:38.321945 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon-log" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.321951 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon-log" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.322123 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="sg-core" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.322140 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-notification-agent" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.322152 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="ceilometer-central-agent" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.322166 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon-log" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.322180 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" containerName="horizon" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.322194 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" containerName="proxy-httpd" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.324032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.327074 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.330162 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.341608 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.462127 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.462194 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glmm4\" (UniqueName: \"kubernetes.io/projected/e9bb2a36-163b-4637-a653-1bdee82ce1d1-kube-api-access-glmm4\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.462227 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-log-httpd\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.462264 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-config-data\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.462308 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-run-httpd\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.462337 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.462355 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-scripts\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564319 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-run-httpd\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564410 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564449 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-scripts\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564507 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564561 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glmm4\" (UniqueName: \"kubernetes.io/projected/e9bb2a36-163b-4637-a653-1bdee82ce1d1-kube-api-access-glmm4\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564596 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-log-httpd\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564640 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-config-data\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.564969 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-run-httpd\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.565446 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-log-httpd\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.571149 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.571193 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-config-data\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.573571 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-scripts\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.580890 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.587025 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glmm4\" (UniqueName: \"kubernetes.io/projected/e9bb2a36-163b-4637-a653-1bdee82ce1d1-kube-api-access-glmm4\") pod \"ceilometer-0\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.651995 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.682731 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:38 crc kubenswrapper[4894]: I1208 15:05:38.687146 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-567d6f4bf7-xrnjd" Dec 08 15:05:39 crc kubenswrapper[4894]: I1208 15:05:39.185632 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:39 crc kubenswrapper[4894]: I1208 15:05:39.237038 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c952e58-c5cb-494c-89fe-b34c835ba0d5" path="/var/lib/kubelet/pods/9c952e58-c5cb-494c-89fe-b34c835ba0d5/volumes" Dec 08 15:05:39 crc kubenswrapper[4894]: I1208 15:05:39.239231 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a44fcb75-8381-44cf-8c5b-2b1108232807" path="/var/lib/kubelet/pods/a44fcb75-8381-44cf-8c5b-2b1108232807/volumes" Dec 08 15:05:39 crc kubenswrapper[4894]: I1208 15:05:39.259372 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerStarted","Data":"fe8cf8a27bcd97b9bbab0a06ca1aad4d2e7fcad51a56ed7890949e7bcdc76232"} Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.393623 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-7hskq"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.395788 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.404637 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-7hskq"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.480348 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-nf84z"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.481792 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.491515 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nf84z"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.518033 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-operator-scripts\") pod \"nova-api-db-create-7hskq\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.518642 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkgj6\" (UniqueName: \"kubernetes.io/projected/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-kube-api-access-kkgj6\") pod \"nova-api-db-create-7hskq\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.597642 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-2gjd8"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.599317 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.620908 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552bb30e-121a-4de7-ab1f-b9d4ae181482-operator-scripts\") pod \"nova-cell0-db-create-nf84z\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.620989 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-operator-scripts\") pod \"nova-api-db-create-7hskq\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.621027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6pll\" (UniqueName: \"kubernetes.io/projected/552bb30e-121a-4de7-ab1f-b9d4ae181482-kube-api-access-r6pll\") pod \"nova-cell0-db-create-nf84z\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.621053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkgj6\" (UniqueName: \"kubernetes.io/projected/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-kube-api-access-kkgj6\") pod \"nova-api-db-create-7hskq\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.622620 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-operator-scripts\") pod \"nova-api-db-create-7hskq\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.622684 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2gjd8"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.646420 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkgj6\" (UniqueName: \"kubernetes.io/projected/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-kube-api-access-kkgj6\") pod \"nova-api-db-create-7hskq\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.692771 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c9d7-account-create-update-fkrpz"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.693964 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.700795 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.725329 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6pll\" (UniqueName: \"kubernetes.io/projected/552bb30e-121a-4de7-ab1f-b9d4ae181482-kube-api-access-r6pll\") pod \"nova-cell0-db-create-nf84z\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.725405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jprbp\" (UniqueName: \"kubernetes.io/projected/374fd9cd-a481-46ba-80b5-6024953cd68d-kube-api-access-jprbp\") pod \"nova-cell1-db-create-2gjd8\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.725516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/374fd9cd-a481-46ba-80b5-6024953cd68d-operator-scripts\") pod \"nova-cell1-db-create-2gjd8\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.725553 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552bb30e-121a-4de7-ab1f-b9d4ae181482-operator-scripts\") pod \"nova-cell0-db-create-nf84z\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.725924 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c9d7-account-create-update-fkrpz"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.726340 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552bb30e-121a-4de7-ab1f-b9d4ae181482-operator-scripts\") pod \"nova-cell0-db-create-nf84z\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.733640 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.756460 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6pll\" (UniqueName: \"kubernetes.io/projected/552bb30e-121a-4de7-ab1f-b9d4ae181482-kube-api-access-r6pll\") pod \"nova-cell0-db-create-nf84z\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.802413 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.806559 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8c84-account-create-update-vrjkn"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.808253 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.810116 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.827509 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jprbp\" (UniqueName: \"kubernetes.io/projected/374fd9cd-a481-46ba-80b5-6024953cd68d-kube-api-access-jprbp\") pod \"nova-cell1-db-create-2gjd8\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.827613 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p4r7\" (UniqueName: \"kubernetes.io/projected/147a0472-ae65-499b-ad28-c5111f7ba6b2-kube-api-access-7p4r7\") pod \"nova-api-c9d7-account-create-update-fkrpz\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.827654 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/374fd9cd-a481-46ba-80b5-6024953cd68d-operator-scripts\") pod \"nova-cell1-db-create-2gjd8\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.827692 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147a0472-ae65-499b-ad28-c5111f7ba6b2-operator-scripts\") pod \"nova-api-c9d7-account-create-update-fkrpz\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.828860 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/374fd9cd-a481-46ba-80b5-6024953cd68d-operator-scripts\") pod \"nova-cell1-db-create-2gjd8\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.854428 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jprbp\" (UniqueName: \"kubernetes.io/projected/374fd9cd-a481-46ba-80b5-6024953cd68d-kube-api-access-jprbp\") pod \"nova-cell1-db-create-2gjd8\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.886240 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8c84-account-create-update-vrjkn"] Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.901962 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.930082 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p4r7\" (UniqueName: \"kubernetes.io/projected/147a0472-ae65-499b-ad28-c5111f7ba6b2-kube-api-access-7p4r7\") pod \"nova-api-c9d7-account-create-update-fkrpz\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.930143 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0062ed70-9c1e-4201-85bd-9a01061bdfa0-operator-scripts\") pod \"nova-cell0-8c84-account-create-update-vrjkn\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.930183 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbz9z\" (UniqueName: \"kubernetes.io/projected/0062ed70-9c1e-4201-85bd-9a01061bdfa0-kube-api-access-hbz9z\") pod \"nova-cell0-8c84-account-create-update-vrjkn\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.930234 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147a0472-ae65-499b-ad28-c5111f7ba6b2-operator-scripts\") pod \"nova-api-c9d7-account-create-update-fkrpz\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.931299 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147a0472-ae65-499b-ad28-c5111f7ba6b2-operator-scripts\") pod \"nova-api-c9d7-account-create-update-fkrpz\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:40 crc kubenswrapper[4894]: I1208 15:05:40.956342 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p4r7\" (UniqueName: \"kubernetes.io/projected/147a0472-ae65-499b-ad28-c5111f7ba6b2-kube-api-access-7p4r7\") pod \"nova-api-c9d7-account-create-update-fkrpz\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.011460 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-024d-account-create-update-6zh4q"] Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.013426 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.021520 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.029261 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-024d-account-create-update-6zh4q"] Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.031923 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0062ed70-9c1e-4201-85bd-9a01061bdfa0-operator-scripts\") pod \"nova-cell0-8c84-account-create-update-vrjkn\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.031982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbz9z\" (UniqueName: \"kubernetes.io/projected/0062ed70-9c1e-4201-85bd-9a01061bdfa0-kube-api-access-hbz9z\") pod \"nova-cell0-8c84-account-create-update-vrjkn\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.034761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0062ed70-9c1e-4201-85bd-9a01061bdfa0-operator-scripts\") pod \"nova-cell0-8c84-account-create-update-vrjkn\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.055700 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbz9z\" (UniqueName: \"kubernetes.io/projected/0062ed70-9c1e-4201-85bd-9a01061bdfa0-kube-api-access-hbz9z\") pod \"nova-cell0-8c84-account-create-update-vrjkn\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.134068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp44w\" (UniqueName: \"kubernetes.io/projected/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-kube-api-access-dp44w\") pod \"nova-cell1-024d-account-create-update-6zh4q\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.134198 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-operator-scripts\") pod \"nova-cell1-024d-account-create-update-6zh4q\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.210779 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.220370 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.236099 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp44w\" (UniqueName: \"kubernetes.io/projected/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-kube-api-access-dp44w\") pod \"nova-cell1-024d-account-create-update-6zh4q\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.236341 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-operator-scripts\") pod \"nova-cell1-024d-account-create-update-6zh4q\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.237623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-operator-scripts\") pod \"nova-cell1-024d-account-create-update-6zh4q\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.270836 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp44w\" (UniqueName: \"kubernetes.io/projected/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-kube-api-access-dp44w\") pod \"nova-cell1-024d-account-create-update-6zh4q\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.308893 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerStarted","Data":"4171dc4cf8071e88a763da1d485c6b51d42f512db1c6979bcc6891bf881d5a6c"} Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.339167 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-7hskq"] Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.350671 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.455177 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nf84z"] Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.618435 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-2gjd8"] Dec 08 15:05:41 crc kubenswrapper[4894]: W1208 15:05:41.629145 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod374fd9cd_a481_46ba_80b5_6024953cd68d.slice/crio-85ca3a4fbf7ea457d8fcb82e9ffef812642879bde42c84cbb4b4022a064f63f2 WatchSource:0}: Error finding container 85ca3a4fbf7ea457d8fcb82e9ffef812642879bde42c84cbb4b4022a064f63f2: Status 404 returned error can't find the container with id 85ca3a4fbf7ea457d8fcb82e9ffef812642879bde42c84cbb4b4022a064f63f2 Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.791741 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8c84-account-create-update-vrjkn"] Dec 08 15:05:41 crc kubenswrapper[4894]: I1208 15:05:41.882785 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c9d7-account-create-update-fkrpz"] Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.029362 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-024d-account-create-update-6zh4q"] Dec 08 15:05:42 crc kubenswrapper[4894]: W1208 15:05:42.054478 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda868f1eb_3d98_45c7_abaf_775e4f4e7db3.slice/crio-111c4b694ecab0c3560ec31e166e1b6bf11e602fe92a747b2583a27eaf738f95 WatchSource:0}: Error finding container 111c4b694ecab0c3560ec31e166e1b6bf11e602fe92a747b2583a27eaf738f95: Status 404 returned error can't find the container with id 111c4b694ecab0c3560ec31e166e1b6bf11e602fe92a747b2583a27eaf738f95 Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.337308 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2gjd8" event={"ID":"374fd9cd-a481-46ba-80b5-6024953cd68d","Type":"ContainerStarted","Data":"fb7f69ae479eaaa6a65aaf7d990283ba529dcea3cab4d0a8e611f0a100c8de01"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.337854 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2gjd8" event={"ID":"374fd9cd-a481-46ba-80b5-6024953cd68d","Type":"ContainerStarted","Data":"85ca3a4fbf7ea457d8fcb82e9ffef812642879bde42c84cbb4b4022a064f63f2"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.344000 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-024d-account-create-update-6zh4q" event={"ID":"a868f1eb-3d98-45c7-abaf-775e4f4e7db3","Type":"ContainerStarted","Data":"111c4b694ecab0c3560ec31e166e1b6bf11e602fe92a747b2583a27eaf738f95"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.354533 4894 generic.go:334] "Generic (PLEG): container finished" podID="8cceeee5-f000-4e96-a019-ef0f1c1e9b09" containerID="be5cfa049be2274e6588909adc1b360ddc25f2e4106ac48e1f25de455dd749af" exitCode=0 Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.354803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7hskq" event={"ID":"8cceeee5-f000-4e96-a019-ef0f1c1e9b09","Type":"ContainerDied","Data":"be5cfa049be2274e6588909adc1b360ddc25f2e4106ac48e1f25de455dd749af"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.354880 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7hskq" event={"ID":"8cceeee5-f000-4e96-a019-ef0f1c1e9b09","Type":"ContainerStarted","Data":"e9da66df9c178028ffce7d5ea69004cf3c2eab2879e35616132c4eb4e078ea4f"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.362875 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" event={"ID":"0062ed70-9c1e-4201-85bd-9a01061bdfa0","Type":"ContainerStarted","Data":"55e524ca71c981b718fa4cb601f1fc6601348c28640edb9425ce62cb401a1f29"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.362918 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" event={"ID":"0062ed70-9c1e-4201-85bd-9a01061bdfa0","Type":"ContainerStarted","Data":"a846c700a72313a5e348ffecb903f0fb4a1e006ee8bfcb65d167b47d00021559"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.380135 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-2gjd8" podStartSLOduration=2.380107727 podStartE2EDuration="2.380107727s" podCreationTimestamp="2025-12-08 15:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:42.362234307 +0000 UTC m=+1163.462240422" watchObservedRunningTime="2025-12-08 15:05:42.380107727 +0000 UTC m=+1163.480113842" Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.388883 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" event={"ID":"147a0472-ae65-499b-ad28-c5111f7ba6b2","Type":"ContainerStarted","Data":"52b30224bcfff18b6d78842f366c717a5842da23c245ee7ca9c0a6f2ff55dbfb"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.388950 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" event={"ID":"147a0472-ae65-499b-ad28-c5111f7ba6b2","Type":"ContainerStarted","Data":"236cc6286b9f3c5ed34d87b1bea8e299668378ba0afce09cc194b9ba67fc196a"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.397421 4894 generic.go:334] "Generic (PLEG): container finished" podID="552bb30e-121a-4de7-ab1f-b9d4ae181482" containerID="0db2de46b62ee56858e9dccbea460b5bdc4a012d009472831e81dd2f6aeba3d8" exitCode=0 Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.397561 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nf84z" event={"ID":"552bb30e-121a-4de7-ab1f-b9d4ae181482","Type":"ContainerDied","Data":"0db2de46b62ee56858e9dccbea460b5bdc4a012d009472831e81dd2f6aeba3d8"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.397653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nf84z" event={"ID":"552bb30e-121a-4de7-ab1f-b9d4ae181482","Type":"ContainerStarted","Data":"6056efcbd7449d2868a41a34b423019973c3133861c4e9d63e1823c8703d42ee"} Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.415717 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" podStartSLOduration=2.415689201 podStartE2EDuration="2.415689201s" podCreationTimestamp="2025-12-08 15:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:42.407244227 +0000 UTC m=+1163.507250352" watchObservedRunningTime="2025-12-08 15:05:42.415689201 +0000 UTC m=+1163.515695316" Dec 08 15:05:42 crc kubenswrapper[4894]: I1208 15:05:42.455787 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" podStartSLOduration=2.455744546 podStartE2EDuration="2.455744546s" podCreationTimestamp="2025-12-08 15:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:42.427885544 +0000 UTC m=+1163.527891649" watchObservedRunningTime="2025-12-08 15:05:42.455744546 +0000 UTC m=+1163.555750661" Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.413304 4894 generic.go:334] "Generic (PLEG): container finished" podID="0062ed70-9c1e-4201-85bd-9a01061bdfa0" containerID="55e524ca71c981b718fa4cb601f1fc6601348c28640edb9425ce62cb401a1f29" exitCode=0 Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.414186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" event={"ID":"0062ed70-9c1e-4201-85bd-9a01061bdfa0","Type":"ContainerDied","Data":"55e524ca71c981b718fa4cb601f1fc6601348c28640edb9425ce62cb401a1f29"} Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.418506 4894 generic.go:334] "Generic (PLEG): container finished" podID="147a0472-ae65-499b-ad28-c5111f7ba6b2" containerID="52b30224bcfff18b6d78842f366c717a5842da23c245ee7ca9c0a6f2ff55dbfb" exitCode=0 Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.418599 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" event={"ID":"147a0472-ae65-499b-ad28-c5111f7ba6b2","Type":"ContainerDied","Data":"52b30224bcfff18b6d78842f366c717a5842da23c245ee7ca9c0a6f2ff55dbfb"} Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.421071 4894 generic.go:334] "Generic (PLEG): container finished" podID="374fd9cd-a481-46ba-80b5-6024953cd68d" containerID="fb7f69ae479eaaa6a65aaf7d990283ba529dcea3cab4d0a8e611f0a100c8de01" exitCode=0 Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.421367 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2gjd8" event={"ID":"374fd9cd-a481-46ba-80b5-6024953cd68d","Type":"ContainerDied","Data":"fb7f69ae479eaaa6a65aaf7d990283ba529dcea3cab4d0a8e611f0a100c8de01"} Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.424144 4894 generic.go:334] "Generic (PLEG): container finished" podID="a868f1eb-3d98-45c7-abaf-775e4f4e7db3" containerID="8e0eff55444972e050b2d1c575a3a5d7c6de042add02fb2c3953f8fdcdfef3f5" exitCode=0 Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.424272 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-024d-account-create-update-6zh4q" event={"ID":"a868f1eb-3d98-45c7-abaf-775e4f4e7db3","Type":"ContainerDied","Data":"8e0eff55444972e050b2d1c575a3a5d7c6de042add02fb2c3953f8fdcdfef3f5"} Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.427843 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerStarted","Data":"48f5d26e8eb1f0d0f7b37ed8c4bd9631572f2e9576fdb48e4592af110a29f226"} Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.836184 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.943799 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.950326 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.954246 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552bb30e-121a-4de7-ab1f-b9d4ae181482-operator-scripts\") pod \"552bb30e-121a-4de7-ab1f-b9d4ae181482\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.954471 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6pll\" (UniqueName: \"kubernetes.io/projected/552bb30e-121a-4de7-ab1f-b9d4ae181482-kube-api-access-r6pll\") pod \"552bb30e-121a-4de7-ab1f-b9d4ae181482\" (UID: \"552bb30e-121a-4de7-ab1f-b9d4ae181482\") " Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.954741 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/552bb30e-121a-4de7-ab1f-b9d4ae181482-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "552bb30e-121a-4de7-ab1f-b9d4ae181482" (UID: "552bb30e-121a-4de7-ab1f-b9d4ae181482"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.954969 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/552bb30e-121a-4de7-ab1f-b9d4ae181482-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:43 crc kubenswrapper[4894]: I1208 15:05:43.961509 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/552bb30e-121a-4de7-ab1f-b9d4ae181482-kube-api-access-r6pll" (OuterVolumeSpecName: "kube-api-access-r6pll") pod "552bb30e-121a-4de7-ab1f-b9d4ae181482" (UID: "552bb30e-121a-4de7-ab1f-b9d4ae181482"). InnerVolumeSpecName "kube-api-access-r6pll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.056496 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkgj6\" (UniqueName: \"kubernetes.io/projected/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-kube-api-access-kkgj6\") pod \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.057323 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-operator-scripts\") pod \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\" (UID: \"8cceeee5-f000-4e96-a019-ef0f1c1e9b09\") " Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.058343 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8cceeee5-f000-4e96-a019-ef0f1c1e9b09" (UID: "8cceeee5-f000-4e96-a019-ef0f1c1e9b09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.058533 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6pll\" (UniqueName: \"kubernetes.io/projected/552bb30e-121a-4de7-ab1f-b9d4ae181482-kube-api-access-r6pll\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.058552 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.059560 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-kube-api-access-kkgj6" (OuterVolumeSpecName: "kube-api-access-kkgj6") pod "8cceeee5-f000-4e96-a019-ef0f1c1e9b09" (UID: "8cceeee5-f000-4e96-a019-ef0f1c1e9b09"). InnerVolumeSpecName "kube-api-access-kkgj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.161044 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkgj6\" (UniqueName: \"kubernetes.io/projected/8cceeee5-f000-4e96-a019-ef0f1c1e9b09-kube-api-access-kkgj6\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.442214 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-7hskq" event={"ID":"8cceeee5-f000-4e96-a019-ef0f1c1e9b09","Type":"ContainerDied","Data":"e9da66df9c178028ffce7d5ea69004cf3c2eab2879e35616132c4eb4e078ea4f"} Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.442267 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-7hskq" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.442280 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9da66df9c178028ffce7d5ea69004cf3c2eab2879e35616132c4eb4e078ea4f" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.445354 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nf84z" event={"ID":"552bb30e-121a-4de7-ab1f-b9d4ae181482","Type":"ContainerDied","Data":"6056efcbd7449d2868a41a34b423019973c3133861c4e9d63e1823c8703d42ee"} Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.445404 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6056efcbd7449d2868a41a34b423019973c3133861c4e9d63e1823c8703d42ee" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.445416 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nf84z" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.448877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerStarted","Data":"3586d73fb8f3f31ef159c004235e510ca56be6bf82e02f966b8cb51b99c69b74"} Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.948075 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.996398 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147a0472-ae65-499b-ad28-c5111f7ba6b2-operator-scripts\") pod \"147a0472-ae65-499b-ad28-c5111f7ba6b2\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.996854 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p4r7\" (UniqueName: \"kubernetes.io/projected/147a0472-ae65-499b-ad28-c5111f7ba6b2-kube-api-access-7p4r7\") pod \"147a0472-ae65-499b-ad28-c5111f7ba6b2\" (UID: \"147a0472-ae65-499b-ad28-c5111f7ba6b2\") " Dec 08 15:05:44 crc kubenswrapper[4894]: I1208 15:05:44.998133 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147a0472-ae65-499b-ad28-c5111f7ba6b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "147a0472-ae65-499b-ad28-c5111f7ba6b2" (UID: "147a0472-ae65-499b-ad28-c5111f7ba6b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.009044 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147a0472-ae65-499b-ad28-c5111f7ba6b2-kube-api-access-7p4r7" (OuterVolumeSpecName: "kube-api-access-7p4r7") pod "147a0472-ae65-499b-ad28-c5111f7ba6b2" (UID: "147a0472-ae65-499b-ad28-c5111f7ba6b2"). InnerVolumeSpecName "kube-api-access-7p4r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.100041 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/147a0472-ae65-499b-ad28-c5111f7ba6b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.100079 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p4r7\" (UniqueName: \"kubernetes.io/projected/147a0472-ae65-499b-ad28-c5111f7ba6b2-kube-api-access-7p4r7\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.126176 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.132047 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.138532 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.202189 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbz9z\" (UniqueName: \"kubernetes.io/projected/0062ed70-9c1e-4201-85bd-9a01061bdfa0-kube-api-access-hbz9z\") pod \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.202382 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp44w\" (UniqueName: \"kubernetes.io/projected/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-kube-api-access-dp44w\") pod \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.203040 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jprbp\" (UniqueName: \"kubernetes.io/projected/374fd9cd-a481-46ba-80b5-6024953cd68d-kube-api-access-jprbp\") pod \"374fd9cd-a481-46ba-80b5-6024953cd68d\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.203122 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-operator-scripts\") pod \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\" (UID: \"a868f1eb-3d98-45c7-abaf-775e4f4e7db3\") " Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.203249 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0062ed70-9c1e-4201-85bd-9a01061bdfa0-operator-scripts\") pod \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\" (UID: \"0062ed70-9c1e-4201-85bd-9a01061bdfa0\") " Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.203294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/374fd9cd-a481-46ba-80b5-6024953cd68d-operator-scripts\") pod \"374fd9cd-a481-46ba-80b5-6024953cd68d\" (UID: \"374fd9cd-a481-46ba-80b5-6024953cd68d\") " Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.204201 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a868f1eb-3d98-45c7-abaf-775e4f4e7db3" (UID: "a868f1eb-3d98-45c7-abaf-775e4f4e7db3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.204416 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/374fd9cd-a481-46ba-80b5-6024953cd68d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "374fd9cd-a481-46ba-80b5-6024953cd68d" (UID: "374fd9cd-a481-46ba-80b5-6024953cd68d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.205276 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0062ed70-9c1e-4201-85bd-9a01061bdfa0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0062ed70-9c1e-4201-85bd-9a01061bdfa0" (UID: "0062ed70-9c1e-4201-85bd-9a01061bdfa0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.212469 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0062ed70-9c1e-4201-85bd-9a01061bdfa0-kube-api-access-hbz9z" (OuterVolumeSpecName: "kube-api-access-hbz9z") pod "0062ed70-9c1e-4201-85bd-9a01061bdfa0" (UID: "0062ed70-9c1e-4201-85bd-9a01061bdfa0"). InnerVolumeSpecName "kube-api-access-hbz9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.212515 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/374fd9cd-a481-46ba-80b5-6024953cd68d-kube-api-access-jprbp" (OuterVolumeSpecName: "kube-api-access-jprbp") pod "374fd9cd-a481-46ba-80b5-6024953cd68d" (UID: "374fd9cd-a481-46ba-80b5-6024953cd68d"). InnerVolumeSpecName "kube-api-access-jprbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.215575 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-kube-api-access-dp44w" (OuterVolumeSpecName: "kube-api-access-dp44w") pod "a868f1eb-3d98-45c7-abaf-775e4f4e7db3" (UID: "a868f1eb-3d98-45c7-abaf-775e4f4e7db3"). InnerVolumeSpecName "kube-api-access-dp44w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.305596 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jprbp\" (UniqueName: \"kubernetes.io/projected/374fd9cd-a481-46ba-80b5-6024953cd68d-kube-api-access-jprbp\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.305632 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.305642 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0062ed70-9c1e-4201-85bd-9a01061bdfa0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.305650 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/374fd9cd-a481-46ba-80b5-6024953cd68d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.305659 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbz9z\" (UniqueName: \"kubernetes.io/projected/0062ed70-9c1e-4201-85bd-9a01061bdfa0-kube-api-access-hbz9z\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.305668 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp44w\" (UniqueName: \"kubernetes.io/projected/a868f1eb-3d98-45c7-abaf-775e4f4e7db3-kube-api-access-dp44w\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.463388 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-024d-account-create-update-6zh4q" event={"ID":"a868f1eb-3d98-45c7-abaf-775e4f4e7db3","Type":"ContainerDied","Data":"111c4b694ecab0c3560ec31e166e1b6bf11e602fe92a747b2583a27eaf738f95"} Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.463458 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="111c4b694ecab0c3560ec31e166e1b6bf11e602fe92a747b2583a27eaf738f95" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.463544 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-024d-account-create-update-6zh4q" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.486799 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerStarted","Data":"7d066623f52fe2d39e59b19dd6125fef4f6a5bc66e07931958bb40c1f6c49d2d"} Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.488265 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-central-agent" containerID="cri-o://4171dc4cf8071e88a763da1d485c6b51d42f512db1c6979bcc6891bf881d5a6c" gracePeriod=30 Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.488324 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="sg-core" containerID="cri-o://3586d73fb8f3f31ef159c004235e510ca56be6bf82e02f966b8cb51b99c69b74" gracePeriod=30 Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.488379 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="proxy-httpd" containerID="cri-o://7d066623f52fe2d39e59b19dd6125fef4f6a5bc66e07931958bb40c1f6c49d2d" gracePeriod=30 Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.488340 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.488287 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-notification-agent" containerID="cri-o://48f5d26e8eb1f0d0f7b37ed8c4bd9631572f2e9576fdb48e4592af110a29f226" gracePeriod=30 Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.500611 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.500605 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8c84-account-create-update-vrjkn" event={"ID":"0062ed70-9c1e-4201-85bd-9a01061bdfa0","Type":"ContainerDied","Data":"a846c700a72313a5e348ffecb903f0fb4a1e006ee8bfcb65d167b47d00021559"} Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.501827 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a846c700a72313a5e348ffecb903f0fb4a1e006ee8bfcb65d167b47d00021559" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.507558 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" event={"ID":"147a0472-ae65-499b-ad28-c5111f7ba6b2","Type":"ContainerDied","Data":"236cc6286b9f3c5ed34d87b1bea8e299668378ba0afce09cc194b9ba67fc196a"} Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.507621 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="236cc6286b9f3c5ed34d87b1bea8e299668378ba0afce09cc194b9ba67fc196a" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.507766 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c9d7-account-create-update-fkrpz" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.521021 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-2gjd8" event={"ID":"374fd9cd-a481-46ba-80b5-6024953cd68d","Type":"ContainerDied","Data":"85ca3a4fbf7ea457d8fcb82e9ffef812642879bde42c84cbb4b4022a064f63f2"} Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.521550 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ca3a4fbf7ea457d8fcb82e9ffef812642879bde42c84cbb4b4022a064f63f2" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.521148 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-2gjd8" Dec 08 15:05:45 crc kubenswrapper[4894]: I1208 15:05:45.534540 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.031272824 podStartE2EDuration="7.534511189s" podCreationTimestamp="2025-12-08 15:05:38 +0000 UTC" firstStartedPulling="2025-12-08 15:05:39.21786096 +0000 UTC m=+1160.317867075" lastFinishedPulling="2025-12-08 15:05:44.721099325 +0000 UTC m=+1165.821105440" observedRunningTime="2025-12-08 15:05:45.515066871 +0000 UTC m=+1166.615072986" watchObservedRunningTime="2025-12-08 15:05:45.534511189 +0000 UTC m=+1166.634517304" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.056347 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjlmx"] Dec 08 15:05:46 crc kubenswrapper[4894]: E1208 15:05:46.057062 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a868f1eb-3d98-45c7-abaf-775e4f4e7db3" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057304 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a868f1eb-3d98-45c7-abaf-775e4f4e7db3" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: E1208 15:05:46.057316 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cceeee5-f000-4e96-a019-ef0f1c1e9b09" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057323 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cceeee5-f000-4e96-a019-ef0f1c1e9b09" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: E1208 15:05:46.057335 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147a0472-ae65-499b-ad28-c5111f7ba6b2" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057341 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="147a0472-ae65-499b-ad28-c5111f7ba6b2" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: E1208 15:05:46.057358 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0062ed70-9c1e-4201-85bd-9a01061bdfa0" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057364 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0062ed70-9c1e-4201-85bd-9a01061bdfa0" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: E1208 15:05:46.057381 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="374fd9cd-a481-46ba-80b5-6024953cd68d" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057386 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="374fd9cd-a481-46ba-80b5-6024953cd68d" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: E1208 15:05:46.057402 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552bb30e-121a-4de7-ab1f-b9d4ae181482" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057414 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="552bb30e-121a-4de7-ab1f-b9d4ae181482" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057653 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0062ed70-9c1e-4201-85bd-9a01061bdfa0" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057680 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a868f1eb-3d98-45c7-abaf-775e4f4e7db3" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057696 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="374fd9cd-a481-46ba-80b5-6024953cd68d" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057708 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="147a0472-ae65-499b-ad28-c5111f7ba6b2" containerName="mariadb-account-create-update" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057726 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cceeee5-f000-4e96-a019-ef0f1c1e9b09" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.057739 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="552bb30e-121a-4de7-ab1f-b9d4ae181482" containerName="mariadb-database-create" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.058707 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.072763 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-krq5v" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.082375 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.082625 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.094363 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjlmx"] Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.153022 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-scripts\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.153197 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkvcm\" (UniqueName: \"kubernetes.io/projected/6c418af0-bc4d-4fe4-880a-804fd7412ec8-kube-api-access-qkvcm\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.153236 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.153259 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-config-data\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.255357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-config-data\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.255465 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-scripts\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.256783 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkvcm\" (UniqueName: \"kubernetes.io/projected/6c418af0-bc4d-4fe4-880a-804fd7412ec8-kube-api-access-qkvcm\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.256860 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.269111 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-config-data\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.274384 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.286348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-scripts\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.298469 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkvcm\" (UniqueName: \"kubernetes.io/projected/6c418af0-bc4d-4fe4-880a-804fd7412ec8-kube-api-access-qkvcm\") pod \"nova-cell0-conductor-db-sync-pjlmx\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.396061 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.535640 4894 generic.go:334] "Generic (PLEG): container finished" podID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerID="7d066623f52fe2d39e59b19dd6125fef4f6a5bc66e07931958bb40c1f6c49d2d" exitCode=0 Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.535677 4894 generic.go:334] "Generic (PLEG): container finished" podID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerID="3586d73fb8f3f31ef159c004235e510ca56be6bf82e02f966b8cb51b99c69b74" exitCode=2 Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.535685 4894 generic.go:334] "Generic (PLEG): container finished" podID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerID="48f5d26e8eb1f0d0f7b37ed8c4bd9631572f2e9576fdb48e4592af110a29f226" exitCode=0 Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.535687 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerDied","Data":"7d066623f52fe2d39e59b19dd6125fef4f6a5bc66e07931958bb40c1f6c49d2d"} Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.535730 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerDied","Data":"3586d73fb8f3f31ef159c004235e510ca56be6bf82e02f966b8cb51b99c69b74"} Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.535743 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerDied","Data":"48f5d26e8eb1f0d0f7b37ed8c4bd9631572f2e9576fdb48e4592af110a29f226"} Dec 08 15:05:46 crc kubenswrapper[4894]: I1208 15:05:46.936255 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjlmx"] Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.140163 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.140667 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-log" containerID="cri-o://02a88beb76b7aedf12add8d09eea22fcbe2316c0e26a00b526674848370c6231" gracePeriod=30 Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.143757 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-httpd" containerID="cri-o://eab57615ea0b8264dc40b7e61bd46025b147db181ff921e453864db32dae66d8" gracePeriod=30 Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.585280 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" event={"ID":"6c418af0-bc4d-4fe4-880a-804fd7412ec8","Type":"ContainerStarted","Data":"3a80185ade8190ce52a39318cd30f253a22bacb82e5d2e14df17b8b29b07ea21"} Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.591394 4894 generic.go:334] "Generic (PLEG): container finished" podID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerID="02a88beb76b7aedf12add8d09eea22fcbe2316c0e26a00b526674848370c6231" exitCode=143 Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.591489 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27","Type":"ContainerDied","Data":"02a88beb76b7aedf12add8d09eea22fcbe2316c0e26a00b526674848370c6231"} Dec 08 15:05:47 crc kubenswrapper[4894]: E1208 15:05:47.594684 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9bb2a36_163b_4637_a653_1bdee82ce1d1.slice/crio-conmon-4171dc4cf8071e88a763da1d485c6b51d42f512db1c6979bcc6891bf881d5a6c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9bb2a36_163b_4637_a653_1bdee82ce1d1.slice/crio-4171dc4cf8071e88a763da1d485c6b51d42f512db1c6979bcc6891bf881d5a6c.scope\": RecentStats: unable to find data in memory cache]" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.597555 4894 generic.go:334] "Generic (PLEG): container finished" podID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerID="4171dc4cf8071e88a763da1d485c6b51d42f512db1c6979bcc6891bf881d5a6c" exitCode=0 Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.597599 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerDied","Data":"4171dc4cf8071e88a763da1d485c6b51d42f512db1c6979bcc6891bf881d5a6c"} Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.874514 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.895294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-config-data\") pod \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.895334 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-sg-core-conf-yaml\") pod \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.895505 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-scripts\") pod \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.895646 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-combined-ca-bundle\") pod \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.895713 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-run-httpd\") pod \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.895758 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-log-httpd\") pod \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.895777 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glmm4\" (UniqueName: \"kubernetes.io/projected/e9bb2a36-163b-4637-a653-1bdee82ce1d1-kube-api-access-glmm4\") pod \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\" (UID: \"e9bb2a36-163b-4637-a653-1bdee82ce1d1\") " Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.902517 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e9bb2a36-163b-4637-a653-1bdee82ce1d1" (UID: "e9bb2a36-163b-4637-a653-1bdee82ce1d1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.903404 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e9bb2a36-163b-4637-a653-1bdee82ce1d1" (UID: "e9bb2a36-163b-4637-a653-1bdee82ce1d1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.946530 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bb2a36-163b-4637-a653-1bdee82ce1d1-kube-api-access-glmm4" (OuterVolumeSpecName: "kube-api-access-glmm4") pod "e9bb2a36-163b-4637-a653-1bdee82ce1d1" (UID: "e9bb2a36-163b-4637-a653-1bdee82ce1d1"). InnerVolumeSpecName "kube-api-access-glmm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.947014 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-scripts" (OuterVolumeSpecName: "scripts") pod "e9bb2a36-163b-4637-a653-1bdee82ce1d1" (UID: "e9bb2a36-163b-4637-a653-1bdee82ce1d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.999528 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.999555 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.999597 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9bb2a36-163b-4637-a653-1bdee82ce1d1-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:47 crc kubenswrapper[4894]: I1208 15:05:47.999605 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glmm4\" (UniqueName: \"kubernetes.io/projected/e9bb2a36-163b-4637-a653-1bdee82ce1d1-kube-api-access-glmm4\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.015966 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9bb2a36-163b-4637-a653-1bdee82ce1d1" (UID: "e9bb2a36-163b-4637-a653-1bdee82ce1d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.034943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e9bb2a36-163b-4637-a653-1bdee82ce1d1" (UID: "e9bb2a36-163b-4637-a653-1bdee82ce1d1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.079565 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-config-data" (OuterVolumeSpecName: "config-data") pod "e9bb2a36-163b-4637-a653-1bdee82ce1d1" (UID: "e9bb2a36-163b-4637-a653-1bdee82ce1d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.101925 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.101982 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.101993 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9bb2a36-163b-4637-a653-1bdee82ce1d1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.610176 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9bb2a36-163b-4637-a653-1bdee82ce1d1","Type":"ContainerDied","Data":"fe8cf8a27bcd97b9bbab0a06ca1aad4d2e7fcad51a56ed7890949e7bcdc76232"} Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.610268 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.610586 4894 scope.go:117] "RemoveContainer" containerID="7d066623f52fe2d39e59b19dd6125fef4f6a5bc66e07931958bb40c1f6c49d2d" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.641546 4894 scope.go:117] "RemoveContainer" containerID="3586d73fb8f3f31ef159c004235e510ca56be6bf82e02f966b8cb51b99c69b74" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.678405 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.703882 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.714981 4894 scope.go:117] "RemoveContainer" containerID="48f5d26e8eb1f0d0f7b37ed8c4bd9631572f2e9576fdb48e4592af110a29f226" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.722808 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:48 crc kubenswrapper[4894]: E1208 15:05:48.723340 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="proxy-httpd" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723360 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="proxy-httpd" Dec 08 15:05:48 crc kubenswrapper[4894]: E1208 15:05:48.723373 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-central-agent" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723380 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-central-agent" Dec 08 15:05:48 crc kubenswrapper[4894]: E1208 15:05:48.723408 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="sg-core" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723415 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="sg-core" Dec 08 15:05:48 crc kubenswrapper[4894]: E1208 15:05:48.723426 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-notification-agent" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723431 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-notification-agent" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723643 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="proxy-httpd" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723702 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-central-agent" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723714 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="sg-core" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.723728 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" containerName="ceilometer-notification-agent" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.725770 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.728773 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.729031 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.735433 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.755576 4894 scope.go:117] "RemoveContainer" containerID="4171dc4cf8071e88a763da1d485c6b51d42f512db1c6979bcc6891bf881d5a6c" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.818144 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-run-httpd\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.818201 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-scripts\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.818262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzrmp\" (UniqueName: \"kubernetes.io/projected/0e88f797-30fa-4aa7-8434-72757307c086-kube-api-access-vzrmp\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.818364 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.818388 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.818420 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-log-httpd\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.818457 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-config-data\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.920725 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzrmp\" (UniqueName: \"kubernetes.io/projected/0e88f797-30fa-4aa7-8434-72757307c086-kube-api-access-vzrmp\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.920920 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.920949 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.920968 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-log-httpd\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.920997 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-config-data\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.921045 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-run-httpd\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.921076 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-scripts\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.924417 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-log-httpd\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.925842 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-run-httpd\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.928349 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.929530 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.930364 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-config-data\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.942296 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-scripts\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.955741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzrmp\" (UniqueName: \"kubernetes.io/projected/0e88f797-30fa-4aa7-8434-72757307c086-kube-api-access-vzrmp\") pod \"ceilometer-0\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " pod="openstack/ceilometer-0" Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.992422 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:48 crc kubenswrapper[4894]: I1208 15:05:48.993256 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.019978 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.020237 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-log" containerID="cri-o://ec3e156e176ec3b29b8d573109a6e2c1c7478ff41628a3698b819bac263d21f0" gracePeriod=30 Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.020390 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-httpd" containerID="cri-o://45b362505a0bb89c90c563f29cab56658bb7e4405698d6e4eb3a96702676892a" gracePeriod=30 Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.226612 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9bb2a36-163b-4637-a653-1bdee82ce1d1" path="/var/lib/kubelet/pods/e9bb2a36-163b-4637-a653-1bdee82ce1d1/volumes" Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.580473 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:05:49 crc kubenswrapper[4894]: W1208 15:05:49.582601 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e88f797_30fa_4aa7_8434_72757307c086.slice/crio-733584d1d0850e33565878345ff955abb1b99c8652fb62857e6c285a243c75c4 WatchSource:0}: Error finding container 733584d1d0850e33565878345ff955abb1b99c8652fb62857e6c285a243c75c4: Status 404 returned error can't find the container with id 733584d1d0850e33565878345ff955abb1b99c8652fb62857e6c285a243c75c4 Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.629537 4894 generic.go:334] "Generic (PLEG): container finished" podID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerID="ec3e156e176ec3b29b8d573109a6e2c1c7478ff41628a3698b819bac263d21f0" exitCode=143 Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.629598 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be1f50cd-10f7-441e-a140-63713dcafc2d","Type":"ContainerDied","Data":"ec3e156e176ec3b29b8d573109a6e2c1c7478ff41628a3698b819bac263d21f0"} Dec 08 15:05:49 crc kubenswrapper[4894]: I1208 15:05:49.631636 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerStarted","Data":"733584d1d0850e33565878345ff955abb1b99c8652fb62857e6c285a243c75c4"} Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.665844 4894 generic.go:334] "Generic (PLEG): container finished" podID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerID="eab57615ea0b8264dc40b7e61bd46025b147db181ff921e453864db32dae66d8" exitCode=0 Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.666867 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27","Type":"ContainerDied","Data":"eab57615ea0b8264dc40b7e61bd46025b147db181ff921e453864db32dae66d8"} Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.675358 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerStarted","Data":"274b7c661c675a921bc3a18d3008a9578a6b50b96b3989dc752e5e691a8afc41"} Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.772659 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858354 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-combined-ca-bundle\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858433 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh228\" (UniqueName: \"kubernetes.io/projected/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-kube-api-access-gh228\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858498 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-scripts\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858530 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-config-data\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858616 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-logs\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858649 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-httpd-run\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858685 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.858745 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-internal-tls-certs\") pod \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\" (UID: \"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27\") " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.860720 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.861051 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-logs" (OuterVolumeSpecName: "logs") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.864673 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-scripts" (OuterVolumeSpecName: "scripts") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.866619 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-kube-api-access-gh228" (OuterVolumeSpecName: "kube-api-access-gh228") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "kube-api-access-gh228". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.867034 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.907315 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.944755 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-config-data" (OuterVolumeSpecName: "config-data") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.961047 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.961081 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.961097 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.961109 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.961154 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.961168 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.961182 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh228\" (UniqueName: \"kubernetes.io/projected/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-kube-api-access-gh228\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.977942 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" (UID: "41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:50 crc kubenswrapper[4894]: I1208 15:05:50.997050 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.062539 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.062806 4894 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.687480 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerStarted","Data":"a610c2bb1259b7de3dc83510b80c961c9308e95254c44d29742ac23638089f05"} Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.689776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27","Type":"ContainerDied","Data":"b537dfa27b314283be1f07456800575b37d24bba219f33649bc694eecdc42f4b"} Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.689832 4894 scope.go:117] "RemoveContainer" containerID="eab57615ea0b8264dc40b7e61bd46025b147db181ff921e453864db32dae66d8" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.689975 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.724403 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.743198 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.765692 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:05:51 crc kubenswrapper[4894]: E1208 15:05:51.766160 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-log" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.766182 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-log" Dec 08 15:05:51 crc kubenswrapper[4894]: E1208 15:05:51.766209 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-httpd" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.766217 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-httpd" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.766475 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-httpd" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.766507 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" containerName="glance-log" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.767744 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.771353 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.771441 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.775691 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.876778 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.876827 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.876986 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62340785-851a-479a-8c14-3908972a9a9f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.877158 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.877254 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdzps\" (UniqueName: \"kubernetes.io/projected/62340785-851a-479a-8c14-3908972a9a9f-kube-api-access-pdzps\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.877303 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.877387 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62340785-851a-479a-8c14-3908972a9a9f-logs\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.877433 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.978892 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.978963 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdzps\" (UniqueName: \"kubernetes.io/projected/62340785-851a-479a-8c14-3908972a9a9f-kube-api-access-pdzps\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.978994 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979029 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62340785-851a-479a-8c14-3908972a9a9f-logs\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979052 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979106 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979142 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979179 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62340785-851a-479a-8c14-3908972a9a9f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979424 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979654 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62340785-851a-479a-8c14-3908972a9a9f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.979879 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62340785-851a-479a-8c14-3908972a9a9f-logs\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.988678 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.989564 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.989724 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:51 crc kubenswrapper[4894]: I1208 15:05:51.991098 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62340785-851a-479a-8c14-3908972a9a9f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:52 crc kubenswrapper[4894]: I1208 15:05:52.009250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdzps\" (UniqueName: \"kubernetes.io/projected/62340785-851a-479a-8c14-3908972a9a9f-kube-api-access-pdzps\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:52 crc kubenswrapper[4894]: I1208 15:05:52.016491 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"62340785-851a-479a-8c14-3908972a9a9f\") " pod="openstack/glance-default-internal-api-0" Dec 08 15:05:52 crc kubenswrapper[4894]: I1208 15:05:52.110573 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 08 15:05:52 crc kubenswrapper[4894]: I1208 15:05:52.745327 4894 generic.go:334] "Generic (PLEG): container finished" podID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerID="45b362505a0bb89c90c563f29cab56658bb7e4405698d6e4eb3a96702676892a" exitCode=0 Dec 08 15:05:52 crc kubenswrapper[4894]: I1208 15:05:52.745382 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be1f50cd-10f7-441e-a140-63713dcafc2d","Type":"ContainerDied","Data":"45b362505a0bb89c90c563f29cab56658bb7e4405698d6e4eb3a96702676892a"} Dec 08 15:05:53 crc kubenswrapper[4894]: I1208 15:05:53.210782 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27" path="/var/lib/kubelet/pods/41f729f3-48ea-4ffe-b4ec-2ea90fe6ff27/volumes" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.239628 4894 scope.go:117] "RemoveContainer" containerID="02a88beb76b7aedf12add8d09eea22fcbe2316c0e26a00b526674848370c6231" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.407867 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573309 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-combined-ca-bundle\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573412 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-scripts\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573439 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-httpd-run\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573466 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmnt5\" (UniqueName: \"kubernetes.io/projected/be1f50cd-10f7-441e-a140-63713dcafc2d-kube-api-access-wmnt5\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573498 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573546 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-config-data\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573596 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-logs\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.573616 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-public-tls-certs\") pod \"be1f50cd-10f7-441e-a140-63713dcafc2d\" (UID: \"be1f50cd-10f7-441e-a140-63713dcafc2d\") " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.581341 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.582437 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-logs" (OuterVolumeSpecName: "logs") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.584082 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.587525 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-scripts" (OuterVolumeSpecName: "scripts") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.596109 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be1f50cd-10f7-441e-a140-63713dcafc2d-kube-api-access-wmnt5" (OuterVolumeSpecName: "kube-api-access-wmnt5") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "kube-api-access-wmnt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.622648 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.688006 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.689270 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.689307 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.689324 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.689338 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmnt5\" (UniqueName: \"kubernetes.io/projected/be1f50cd-10f7-441e-a140-63713dcafc2d-kube-api-access-wmnt5\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.689377 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.689389 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1f50cd-10f7-441e-a140-63713dcafc2d-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.689403 4894 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.693553 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.725752 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-config-data" (OuterVolumeSpecName: "config-data") pod "be1f50cd-10f7-441e-a140-63713dcafc2d" (UID: "be1f50cd-10f7-441e-a140-63713dcafc2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.729895 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.790172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" event={"ID":"6c418af0-bc4d-4fe4-880a-804fd7412ec8","Type":"ContainerStarted","Data":"f82f44abbb6f6215eca56ec63cd083e5e05ee38f02998d93101a0c3443746003"} Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.792166 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.792199 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1f50cd-10f7-441e-a140-63713dcafc2d-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.795110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"62340785-851a-479a-8c14-3908972a9a9f","Type":"ContainerStarted","Data":"9448ca9241e81cf1726b04e4e80d6ffe19060c6746011c0233cc872b86870a02"} Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.801250 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"be1f50cd-10f7-441e-a140-63713dcafc2d","Type":"ContainerDied","Data":"61aee1f2c5b5dd7f73ee5ee3e12e85e30912b3e9ed28d3adba8981ebab0bcde6"} Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.801322 4894 scope.go:117] "RemoveContainer" containerID="45b362505a0bb89c90c563f29cab56658bb7e4405698d6e4eb3a96702676892a" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.801505 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.814044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerStarted","Data":"e9d6385bac32c049c7102f7d64837cc76fddc64ba7f1d26582fcd76ef5585017"} Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.818503 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" podStartSLOduration=1.505346444 podStartE2EDuration="10.818479811s" podCreationTimestamp="2025-12-08 15:05:46 +0000 UTC" firstStartedPulling="2025-12-08 15:05:46.949363341 +0000 UTC m=+1168.049369456" lastFinishedPulling="2025-12-08 15:05:56.262496708 +0000 UTC m=+1177.362502823" observedRunningTime="2025-12-08 15:05:56.804326098 +0000 UTC m=+1177.904332213" watchObservedRunningTime="2025-12-08 15:05:56.818479811 +0000 UTC m=+1177.918485926" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.893061 4894 scope.go:117] "RemoveContainer" containerID="ec3e156e176ec3b29b8d573109a6e2c1c7478ff41628a3698b819bac263d21f0" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.896487 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.912294 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.933910 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:05:56 crc kubenswrapper[4894]: E1208 15:05:56.934366 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-log" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.934381 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-log" Dec 08 15:05:56 crc kubenswrapper[4894]: E1208 15:05:56.934402 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-httpd" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.934411 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-httpd" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.934686 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-log" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.934717 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" containerName="glance-httpd" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.935962 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.944711 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.945526 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 08 15:05:56 crc kubenswrapper[4894]: I1208 15:05:56.947286 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.099665 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.099740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.099760 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2pl9\" (UniqueName: \"kubernetes.io/projected/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-kube-api-access-k2pl9\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.100045 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-logs\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.100128 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.100411 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.100457 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.100542 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203453 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203502 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2pl9\" (UniqueName: \"kubernetes.io/projected/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-kube-api-access-k2pl9\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203526 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-logs\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203552 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203635 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.203658 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.204228 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.204949 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-logs\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.205347 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.208739 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.212653 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.213115 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-config-data\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.219017 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be1f50cd-10f7-441e-a140-63713dcafc2d" path="/var/lib/kubelet/pods/be1f50cd-10f7-441e-a140-63713dcafc2d/volumes" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.226017 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2pl9\" (UniqueName: \"kubernetes.io/projected/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-kube-api-access-k2pl9\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.226054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d3952ab-f60a-4450-bd40-27b2b99a3e9e-scripts\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.272715 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"2d3952ab-f60a-4450-bd40-27b2b99a3e9e\") " pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.568443 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 08 15:05:57 crc kubenswrapper[4894]: I1208 15:05:57.882851 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"62340785-851a-479a-8c14-3908972a9a9f","Type":"ContainerStarted","Data":"4381b4d2c3072a3843409aaecefaf4f6f7da1b11f8170a1b57d9af8fa6117ddf"} Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.180778 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 08 15:05:58 crc kubenswrapper[4894]: W1208 15:05:58.182603 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d3952ab_f60a_4450_bd40_27b2b99a3e9e.slice/crio-7a00d5ef362aecd6361a03b8c59150611e8b7beb0843120307fa0101910f8964 WatchSource:0}: Error finding container 7a00d5ef362aecd6361a03b8c59150611e8b7beb0843120307fa0101910f8964: Status 404 returned error can't find the container with id 7a00d5ef362aecd6361a03b8c59150611e8b7beb0843120307fa0101910f8964 Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.894137 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"62340785-851a-479a-8c14-3908972a9a9f","Type":"ContainerStarted","Data":"b6e0f42c76fde3a7c37865f938769cfbf16de9377453d8fbf3ab8e49cd2fc796"} Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.896721 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2d3952ab-f60a-4450-bd40-27b2b99a3e9e","Type":"ContainerStarted","Data":"db86f0b0f27df83cb09ce154d219e225f4f74ea4b5dfa6be7fc7b2c59151b1bd"} Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.896763 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2d3952ab-f60a-4450-bd40-27b2b99a3e9e","Type":"ContainerStarted","Data":"7a00d5ef362aecd6361a03b8c59150611e8b7beb0843120307fa0101910f8964"} Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.899139 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerStarted","Data":"a5de30d8ee82eda31a4dfcf3131e61c9e9f7cda63f16122fc505fa1a95dea11a"} Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.899315 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-central-agent" containerID="cri-o://274b7c661c675a921bc3a18d3008a9578a6b50b96b3989dc752e5e691a8afc41" gracePeriod=30 Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.899669 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.899735 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="proxy-httpd" containerID="cri-o://a5de30d8ee82eda31a4dfcf3131e61c9e9f7cda63f16122fc505fa1a95dea11a" gracePeriod=30 Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.899807 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="sg-core" containerID="cri-o://e9d6385bac32c049c7102f7d64837cc76fddc64ba7f1d26582fcd76ef5585017" gracePeriod=30 Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.899890 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-notification-agent" containerID="cri-o://a610c2bb1259b7de3dc83510b80c961c9308e95254c44d29742ac23638089f05" gracePeriod=30 Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.931270 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.93124716 podStartE2EDuration="7.93124716s" podCreationTimestamp="2025-12-08 15:05:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:58.924291642 +0000 UTC m=+1180.024297777" watchObservedRunningTime="2025-12-08 15:05:58.93124716 +0000 UTC m=+1180.031253275" Dec 08 15:05:58 crc kubenswrapper[4894]: I1208 15:05:58.953254 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.671601167 podStartE2EDuration="10.953235129s" podCreationTimestamp="2025-12-08 15:05:48 +0000 UTC" firstStartedPulling="2025-12-08 15:05:49.584882902 +0000 UTC m=+1170.684889007" lastFinishedPulling="2025-12-08 15:05:57.866516844 +0000 UTC m=+1178.966522969" observedRunningTime="2025-12-08 15:05:58.948528632 +0000 UTC m=+1180.048534747" watchObservedRunningTime="2025-12-08 15:05:58.953235129 +0000 UTC m=+1180.053241244" Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.910482 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2d3952ab-f60a-4450-bd40-27b2b99a3e9e","Type":"ContainerStarted","Data":"47d8442e79a1818b57130a48ead08186fe4579802ee7813639be1d2450e6a0e6"} Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917018 4894 generic.go:334] "Generic (PLEG): container finished" podID="0e88f797-30fa-4aa7-8434-72757307c086" containerID="a5de30d8ee82eda31a4dfcf3131e61c9e9f7cda63f16122fc505fa1a95dea11a" exitCode=0 Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917284 4894 generic.go:334] "Generic (PLEG): container finished" podID="0e88f797-30fa-4aa7-8434-72757307c086" containerID="e9d6385bac32c049c7102f7d64837cc76fddc64ba7f1d26582fcd76ef5585017" exitCode=2 Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917362 4894 generic.go:334] "Generic (PLEG): container finished" podID="0e88f797-30fa-4aa7-8434-72757307c086" containerID="a610c2bb1259b7de3dc83510b80c961c9308e95254c44d29742ac23638089f05" exitCode=0 Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917511 4894 generic.go:334] "Generic (PLEG): container finished" podID="0e88f797-30fa-4aa7-8434-72757307c086" containerID="274b7c661c675a921bc3a18d3008a9578a6b50b96b3989dc752e5e691a8afc41" exitCode=0 Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917783 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerDied","Data":"a5de30d8ee82eda31a4dfcf3131e61c9e9f7cda63f16122fc505fa1a95dea11a"} Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917856 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerDied","Data":"e9d6385bac32c049c7102f7d64837cc76fddc64ba7f1d26582fcd76ef5585017"} Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917868 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerDied","Data":"a610c2bb1259b7de3dc83510b80c961c9308e95254c44d29742ac23638089f05"} Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917878 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerDied","Data":"274b7c661c675a921bc3a18d3008a9578a6b50b96b3989dc752e5e691a8afc41"} Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917886 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0e88f797-30fa-4aa7-8434-72757307c086","Type":"ContainerDied","Data":"733584d1d0850e33565878345ff955abb1b99c8652fb62857e6c285a243c75c4"} Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.917895 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="733584d1d0850e33565878345ff955abb1b99c8652fb62857e6c285a243c75c4" Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.919471 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:05:59 crc kubenswrapper[4894]: I1208 15:05:59.940329 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.940310793 podStartE2EDuration="3.940310793s" podCreationTimestamp="2025-12-08 15:05:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:05:59.933691946 +0000 UTC m=+1181.033698071" watchObservedRunningTime="2025-12-08 15:05:59.940310793 +0000 UTC m=+1181.040316908" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.071597 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-combined-ca-bundle\") pod \"0e88f797-30fa-4aa7-8434-72757307c086\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.071669 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-run-httpd\") pod \"0e88f797-30fa-4aa7-8434-72757307c086\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.071740 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-scripts\") pod \"0e88f797-30fa-4aa7-8434-72757307c086\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.071789 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-config-data\") pod \"0e88f797-30fa-4aa7-8434-72757307c086\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.072035 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0e88f797-30fa-4aa7-8434-72757307c086" (UID: "0e88f797-30fa-4aa7-8434-72757307c086"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.072462 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-log-httpd\") pod \"0e88f797-30fa-4aa7-8434-72757307c086\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.072958 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzrmp\" (UniqueName: \"kubernetes.io/projected/0e88f797-30fa-4aa7-8434-72757307c086-kube-api-access-vzrmp\") pod \"0e88f797-30fa-4aa7-8434-72757307c086\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.072793 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0e88f797-30fa-4aa7-8434-72757307c086" (UID: "0e88f797-30fa-4aa7-8434-72757307c086"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.073110 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-sg-core-conf-yaml\") pod \"0e88f797-30fa-4aa7-8434-72757307c086\" (UID: \"0e88f797-30fa-4aa7-8434-72757307c086\") " Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.074768 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.074792 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e88f797-30fa-4aa7-8434-72757307c086-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.077933 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-scripts" (OuterVolumeSpecName: "scripts") pod "0e88f797-30fa-4aa7-8434-72757307c086" (UID: "0e88f797-30fa-4aa7-8434-72757307c086"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.078121 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e88f797-30fa-4aa7-8434-72757307c086-kube-api-access-vzrmp" (OuterVolumeSpecName: "kube-api-access-vzrmp") pod "0e88f797-30fa-4aa7-8434-72757307c086" (UID: "0e88f797-30fa-4aa7-8434-72757307c086"). InnerVolumeSpecName "kube-api-access-vzrmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.108971 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0e88f797-30fa-4aa7-8434-72757307c086" (UID: "0e88f797-30fa-4aa7-8434-72757307c086"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.149580 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e88f797-30fa-4aa7-8434-72757307c086" (UID: "0e88f797-30fa-4aa7-8434-72757307c086"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.177014 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.177052 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.177066 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.177079 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzrmp\" (UniqueName: \"kubernetes.io/projected/0e88f797-30fa-4aa7-8434-72757307c086-kube-api-access-vzrmp\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.189735 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-config-data" (OuterVolumeSpecName: "config-data") pod "0e88f797-30fa-4aa7-8434-72757307c086" (UID: "0e88f797-30fa-4aa7-8434-72757307c086"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.279192 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e88f797-30fa-4aa7-8434-72757307c086-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.927743 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.968211 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.976761 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.998683 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:00 crc kubenswrapper[4894]: E1208 15:06:00.999068 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-central-agent" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999082 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-central-agent" Dec 08 15:06:00 crc kubenswrapper[4894]: E1208 15:06:00.999104 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="sg-core" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999111 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="sg-core" Dec 08 15:06:00 crc kubenswrapper[4894]: E1208 15:06:00.999120 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="proxy-httpd" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999126 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="proxy-httpd" Dec 08 15:06:00 crc kubenswrapper[4894]: E1208 15:06:00.999149 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-notification-agent" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999155 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-notification-agent" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999318 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-central-agent" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999337 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="sg-core" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999350 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="ceilometer-notification-agent" Dec 08 15:06:00 crc kubenswrapper[4894]: I1208 15:06:00.999360 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e88f797-30fa-4aa7-8434-72757307c086" containerName="proxy-httpd" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.000953 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.003320 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.003481 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.023231 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.092125 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.092253 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-config-data\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.092299 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-scripts\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.092346 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcpcf\" (UniqueName: \"kubernetes.io/projected/081d95e0-0092-43e3-9517-43a8a5b62d52-kube-api-access-gcpcf\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.092498 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.092544 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-run-httpd\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.092875 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-log-httpd\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.195108 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-log-httpd\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.195165 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.195191 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-config-data\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.195209 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-scripts\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.195232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcpcf\" (UniqueName: \"kubernetes.io/projected/081d95e0-0092-43e3-9517-43a8a5b62d52-kube-api-access-gcpcf\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.195744 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-log-httpd\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.196144 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.196192 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-run-httpd\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.196552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-run-httpd\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.202429 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.203293 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-config-data\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.203950 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.204781 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-scripts\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.207857 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e88f797-30fa-4aa7-8434-72757307c086" path="/var/lib/kubelet/pods/0e88f797-30fa-4aa7-8434-72757307c086/volumes" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.214779 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcpcf\" (UniqueName: \"kubernetes.io/projected/081d95e0-0092-43e3-9517-43a8a5b62d52-kube-api-access-gcpcf\") pod \"ceilometer-0\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.326890 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.757923 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:01 crc kubenswrapper[4894]: W1208 15:06:01.761132 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081d95e0_0092_43e3_9517_43a8a5b62d52.slice/crio-d1662c378fb52cdbe5b191f96a5c91dae438cfb61b0cc9251b704207a00ed34d WatchSource:0}: Error finding container d1662c378fb52cdbe5b191f96a5c91dae438cfb61b0cc9251b704207a00ed34d: Status 404 returned error can't find the container with id d1662c378fb52cdbe5b191f96a5c91dae438cfb61b0cc9251b704207a00ed34d Dec 08 15:06:01 crc kubenswrapper[4894]: I1208 15:06:01.937527 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerStarted","Data":"d1662c378fb52cdbe5b191f96a5c91dae438cfb61b0cc9251b704207a00ed34d"} Dec 08 15:06:02 crc kubenswrapper[4894]: I1208 15:06:02.110972 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:02 crc kubenswrapper[4894]: I1208 15:06:02.111019 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:02 crc kubenswrapper[4894]: I1208 15:06:02.142435 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:02 crc kubenswrapper[4894]: I1208 15:06:02.154968 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:02 crc kubenswrapper[4894]: I1208 15:06:02.949989 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerStarted","Data":"1efe09e6136fe6cd9c2fa9145b6e5de05a578f944d900052be192e81e2e268f4"} Dec 08 15:06:02 crc kubenswrapper[4894]: I1208 15:06:02.950346 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:02 crc kubenswrapper[4894]: I1208 15:06:02.950362 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:03 crc kubenswrapper[4894]: I1208 15:06:03.962594 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerStarted","Data":"e04d41dddd19aca575f13d7970b6b84068d355c9542d66c26cf40079d86c3c57"} Dec 08 15:06:03 crc kubenswrapper[4894]: I1208 15:06:03.962947 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerStarted","Data":"4ffcb97f784d42b5dab0f78f0d8f43a46304afac4c73a3419b869d2dc795f47f"} Dec 08 15:06:04 crc kubenswrapper[4894]: I1208 15:06:04.975115 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerStarted","Data":"2d8a3b0738fa1d5e7b6049581bbb6bac10282d01d55139b385b2bc0b59ec069e"} Dec 08 15:06:04 crc kubenswrapper[4894]: I1208 15:06:04.975491 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:06:04 crc kubenswrapper[4894]: I1208 15:06:04.998705 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.170612573 podStartE2EDuration="4.998680665s" podCreationTimestamp="2025-12-08 15:06:00 +0000 UTC" firstStartedPulling="2025-12-08 15:06:01.763362568 +0000 UTC m=+1182.863368683" lastFinishedPulling="2025-12-08 15:06:04.59143066 +0000 UTC m=+1185.691436775" observedRunningTime="2025-12-08 15:06:04.993697509 +0000 UTC m=+1186.093703624" watchObservedRunningTime="2025-12-08 15:06:04.998680665 +0000 UTC m=+1186.098686780" Dec 08 15:06:05 crc kubenswrapper[4894]: I1208 15:06:05.130415 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:06 crc kubenswrapper[4894]: I1208 15:06:06.023918 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.296839 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.297187 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.297241 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.298119 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9002401f278cd2ddc27524b106a2357f8f007e534b8798b7fd2145f577807600"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.298183 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://9002401f278cd2ddc27524b106a2357f8f007e534b8798b7fd2145f577807600" gracePeriod=600 Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.569698 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.570388 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.604637 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 15:06:07 crc kubenswrapper[4894]: I1208 15:06:07.616029 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 08 15:06:08 crc kubenswrapper[4894]: I1208 15:06:08.014348 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="9002401f278cd2ddc27524b106a2357f8f007e534b8798b7fd2145f577807600" exitCode=0 Dec 08 15:06:08 crc kubenswrapper[4894]: I1208 15:06:08.015590 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"9002401f278cd2ddc27524b106a2357f8f007e534b8798b7fd2145f577807600"} Dec 08 15:06:08 crc kubenswrapper[4894]: I1208 15:06:08.015768 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 15:06:08 crc kubenswrapper[4894]: I1208 15:06:08.015801 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 08 15:06:08 crc kubenswrapper[4894]: I1208 15:06:08.015830 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"d9103dd6f195ccf766c15c1a9fa81340683956ac05b3d32c6e73e01c28e7fa11"} Dec 08 15:06:08 crc kubenswrapper[4894]: I1208 15:06:08.015866 4894 scope.go:117] "RemoveContainer" containerID="1b3439967813e9136ef0135406ae68191aab690f31fec7c75b03962a836112d1" Dec 08 15:06:10 crc kubenswrapper[4894]: I1208 15:06:10.041807 4894 generic.go:334] "Generic (PLEG): container finished" podID="6c418af0-bc4d-4fe4-880a-804fd7412ec8" containerID="f82f44abbb6f6215eca56ec63cd083e5e05ee38f02998d93101a0c3443746003" exitCode=0 Dec 08 15:06:10 crc kubenswrapper[4894]: I1208 15:06:10.041884 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" event={"ID":"6c418af0-bc4d-4fe4-880a-804fd7412ec8","Type":"ContainerDied","Data":"f82f44abbb6f6215eca56ec63cd083e5e05ee38f02998d93101a0c3443746003"} Dec 08 15:06:10 crc kubenswrapper[4894]: I1208 15:06:10.042555 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 15:06:10 crc kubenswrapper[4894]: I1208 15:06:10.042571 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 08 15:06:10 crc kubenswrapper[4894]: I1208 15:06:10.086204 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 15:06:10 crc kubenswrapper[4894]: I1208 15:06:10.128526 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.443861 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.609716 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkvcm\" (UniqueName: \"kubernetes.io/projected/6c418af0-bc4d-4fe4-880a-804fd7412ec8-kube-api-access-qkvcm\") pod \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.610126 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-config-data\") pod \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.610180 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-scripts\") pod \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.610257 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-combined-ca-bundle\") pod \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\" (UID: \"6c418af0-bc4d-4fe4-880a-804fd7412ec8\") " Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.618980 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-scripts" (OuterVolumeSpecName: "scripts") pod "6c418af0-bc4d-4fe4-880a-804fd7412ec8" (UID: "6c418af0-bc4d-4fe4-880a-804fd7412ec8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.633214 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c418af0-bc4d-4fe4-880a-804fd7412ec8-kube-api-access-qkvcm" (OuterVolumeSpecName: "kube-api-access-qkvcm") pod "6c418af0-bc4d-4fe4-880a-804fd7412ec8" (UID: "6c418af0-bc4d-4fe4-880a-804fd7412ec8"). InnerVolumeSpecName "kube-api-access-qkvcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.639964 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-config-data" (OuterVolumeSpecName: "config-data") pod "6c418af0-bc4d-4fe4-880a-804fd7412ec8" (UID: "6c418af0-bc4d-4fe4-880a-804fd7412ec8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.651249 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c418af0-bc4d-4fe4-880a-804fd7412ec8" (UID: "6c418af0-bc4d-4fe4-880a-804fd7412ec8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.713163 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkvcm\" (UniqueName: \"kubernetes.io/projected/6c418af0-bc4d-4fe4-880a-804fd7412ec8-kube-api-access-qkvcm\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.713199 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.713211 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:11 crc kubenswrapper[4894]: I1208 15:06:11.713222 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c418af0-bc4d-4fe4-880a-804fd7412ec8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.060169 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" event={"ID":"6c418af0-bc4d-4fe4-880a-804fd7412ec8","Type":"ContainerDied","Data":"3a80185ade8190ce52a39318cd30f253a22bacb82e5d2e14df17b8b29b07ea21"} Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.060215 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a80185ade8190ce52a39318cd30f253a22bacb82e5d2e14df17b8b29b07ea21" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.060212 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-pjlmx" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.242768 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 08 15:06:12 crc kubenswrapper[4894]: E1208 15:06:12.243166 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c418af0-bc4d-4fe4-880a-804fd7412ec8" containerName="nova-cell0-conductor-db-sync" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.243182 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c418af0-bc4d-4fe4-880a-804fd7412ec8" containerName="nova-cell0-conductor-db-sync" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.243389 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c418af0-bc4d-4fe4-880a-804fd7412ec8" containerName="nova-cell0-conductor-db-sync" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.244027 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.245355 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-krq5v" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.246685 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.252537 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.323739 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96255803-4336-4b64-8aee-956302d0c5a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.323900 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96255803-4336-4b64-8aee-956302d0c5a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.323969 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpzb\" (UniqueName: \"kubernetes.io/projected/96255803-4336-4b64-8aee-956302d0c5a2-kube-api-access-xmpzb\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.425662 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96255803-4336-4b64-8aee-956302d0c5a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.426250 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpzb\" (UniqueName: \"kubernetes.io/projected/96255803-4336-4b64-8aee-956302d0c5a2-kube-api-access-xmpzb\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.426369 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96255803-4336-4b64-8aee-956302d0c5a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.432262 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96255803-4336-4b64-8aee-956302d0c5a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.432313 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96255803-4336-4b64-8aee-956302d0c5a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.446131 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpzb\" (UniqueName: \"kubernetes.io/projected/96255803-4336-4b64-8aee-956302d0c5a2-kube-api-access-xmpzb\") pod \"nova-cell0-conductor-0\" (UID: \"96255803-4336-4b64-8aee-956302d0c5a2\") " pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:12 crc kubenswrapper[4894]: I1208 15:06:12.580444 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:13 crc kubenswrapper[4894]: I1208 15:06:13.003276 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 08 15:06:13 crc kubenswrapper[4894]: W1208 15:06:13.006997 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96255803_4336_4b64_8aee_956302d0c5a2.slice/crio-0a326f23080bb2e2b32500df22a2a7b1f0cab389f9b5bca1b2d23495c798cc64 WatchSource:0}: Error finding container 0a326f23080bb2e2b32500df22a2a7b1f0cab389f9b5bca1b2d23495c798cc64: Status 404 returned error can't find the container with id 0a326f23080bb2e2b32500df22a2a7b1f0cab389f9b5bca1b2d23495c798cc64 Dec 08 15:06:13 crc kubenswrapper[4894]: I1208 15:06:13.075956 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"96255803-4336-4b64-8aee-956302d0c5a2","Type":"ContainerStarted","Data":"0a326f23080bb2e2b32500df22a2a7b1f0cab389f9b5bca1b2d23495c798cc64"} Dec 08 15:06:14 crc kubenswrapper[4894]: I1208 15:06:14.085503 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"96255803-4336-4b64-8aee-956302d0c5a2","Type":"ContainerStarted","Data":"411c792d35c3af4a6bfa03c2d543fdccbcd40461a4d5fe9b04d352b73582ae80"} Dec 08 15:06:14 crc kubenswrapper[4894]: I1208 15:06:14.085675 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:14 crc kubenswrapper[4894]: I1208 15:06:14.109904 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.109882847 podStartE2EDuration="2.109882847s" podCreationTimestamp="2025-12-08 15:06:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:14.104036503 +0000 UTC m=+1195.204042618" watchObservedRunningTime="2025-12-08 15:06:14.109882847 +0000 UTC m=+1195.209888962" Dec 08 15:06:22 crc kubenswrapper[4894]: I1208 15:06:22.610849 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.186430 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-mjdxz"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.187951 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.192509 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.193837 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.216619 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mjdxz"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.322777 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.322855 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wvqk\" (UniqueName: \"kubernetes.io/projected/06425613-4821-4a3d-9507-f6176518464b-kube-api-access-6wvqk\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.323017 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-config-data\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.323035 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-scripts\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.380972 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.382528 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.386469 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.402167 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.419512 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.421436 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.424802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-config-data\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.424884 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-scripts\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.424954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.424995 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wvqk\" (UniqueName: \"kubernetes.io/projected/06425613-4821-4a3d-9507-f6176518464b-kube-api-access-6wvqk\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.426217 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.433691 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-config-data\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.442134 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.446408 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-scripts\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.470931 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.496018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wvqk\" (UniqueName: \"kubernetes.io/projected/06425613-4821-4a3d-9507-f6176518464b-kube-api-access-6wvqk\") pod \"nova-cell0-cell-mapping-mjdxz\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.519248 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.539941 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.540572 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftlp2\" (UniqueName: \"kubernetes.io/projected/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-kube-api-access-ftlp2\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.540762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-logs\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.540893 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.541118 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.541240 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndk5c\" (UniqueName: \"kubernetes.io/projected/8228dc10-46b5-4855-a46c-aa6e16d09735-kube-api-access-ndk5c\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.541371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-config-data\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.569503 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.573180 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.581329 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.635888 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.647985 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-config-data\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.648089 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.648138 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftlp2\" (UniqueName: \"kubernetes.io/projected/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-kube-api-access-ftlp2\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.648199 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-logs\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.648228 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.648321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.648337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndk5c\" (UniqueName: \"kubernetes.io/projected/8228dc10-46b5-4855-a46c-aa6e16d09735-kube-api-access-ndk5c\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.653364 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-logs\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.659323 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-w87xf"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.661942 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.664274 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.664976 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.674115 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-config-data\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.688601 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.688978 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-w87xf"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.700041 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndk5c\" (UniqueName: \"kubernetes.io/projected/8228dc10-46b5-4855-a46c-aa6e16d09735-kube-api-access-ndk5c\") pod \"nova-cell1-novncproxy-0\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.730405 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftlp2\" (UniqueName: \"kubernetes.io/projected/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-kube-api-access-ftlp2\") pod \"nova-metadata-0\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.739283 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.743989 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.747402 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752220 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86r6v\" (UniqueName: \"kubernetes.io/projected/612e0da4-aaaa-4050-be6d-ad14bae102be-kube-api-access-86r6v\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752285 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752323 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752406 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752438 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-config\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752543 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdn2n\" (UniqueName: \"kubernetes.io/projected/0c18cbf5-83bf-424d-9870-1f921b25a9e2-kube-api-access-sdn2n\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752589 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/612e0da4-aaaa-4050-be6d-ad14bae102be-logs\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.752648 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.760402 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.778015 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25csl\" (UniqueName: \"kubernetes.io/projected/3ca71594-fdd4-4d9b-93b2-81f739476f9d-kube-api-access-25csl\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855377 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/612e0da4-aaaa-4050-be6d-ad14bae102be-logs\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855483 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855516 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86r6v\" (UniqueName: \"kubernetes.io/projected/612e0da4-aaaa-4050-be6d-ad14bae102be-kube-api-access-86r6v\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855541 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855585 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855615 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-config-data\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855666 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855745 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.855779 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-config\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.857015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/612e0da4-aaaa-4050-be6d-ad14bae102be-logs\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.857114 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.857150 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.857191 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdn2n\" (UniqueName: \"kubernetes.io/projected/0c18cbf5-83bf-424d-9870-1f921b25a9e2-kube-api-access-sdn2n\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.857678 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.858041 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.859010 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.859377 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.862525 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.863016 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-config\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.871362 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.873780 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86r6v\" (UniqueName: \"kubernetes.io/projected/612e0da4-aaaa-4050-be6d-ad14bae102be-kube-api-access-86r6v\") pod \"nova-api-0\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " pod="openstack/nova-api-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.881484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdn2n\" (UniqueName: \"kubernetes.io/projected/0c18cbf5-83bf-424d-9870-1f921b25a9e2-kube-api-access-sdn2n\") pod \"dnsmasq-dns-845d6d6f59-w87xf\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.959794 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25csl\" (UniqueName: \"kubernetes.io/projected/3ca71594-fdd4-4d9b-93b2-81f739476f9d-kube-api-access-25csl\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.960150 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.960251 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-config-data\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.965780 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.967253 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-config-data\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.981234 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25csl\" (UniqueName: \"kubernetes.io/projected/3ca71594-fdd4-4d9b-93b2-81f739476f9d-kube-api-access-25csl\") pod \"nova-scheduler-0\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:23 crc kubenswrapper[4894]: I1208 15:06:23.999266 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.067568 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.087989 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.115115 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.238957 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mjdxz"] Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.330147 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:24 crc kubenswrapper[4894]: W1208 15:06:24.337991 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dcc0f5d_9f90_44ab_9bc9_f7deb5c8fa31.slice/crio-bc2b143f57fa7cbefbbd6b7dbae10b066c9e4d8c3efb44d2ce92d1d5714e6891 WatchSource:0}: Error finding container bc2b143f57fa7cbefbbd6b7dbae10b066c9e4d8c3efb44d2ce92d1d5714e6891: Status 404 returned error can't find the container with id bc2b143f57fa7cbefbbd6b7dbae10b066c9e4d8c3efb44d2ce92d1d5714e6891 Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.429066 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2b5qb"] Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.431965 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.447182 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.447498 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.483524 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2b5qb"] Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.536211 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.578391 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.578533 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-scripts\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.578742 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nkr5\" (UniqueName: \"kubernetes.io/projected/d73ecea1-981c-4b0a-91f1-ace506d48483-kube-api-access-6nkr5\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.578917 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-config-data\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.641538 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:24 crc kubenswrapper[4894]: W1208 15:06:24.646237 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod612e0da4_aaaa_4050_be6d_ad14bae102be.slice/crio-e28407ed1fb3661f2d151147a973905b6a18fc09cf46e284f9c3bdb39909b354 WatchSource:0}: Error finding container e28407ed1fb3661f2d151147a973905b6a18fc09cf46e284f9c3bdb39909b354: Status 404 returned error can't find the container with id e28407ed1fb3661f2d151147a973905b6a18fc09cf46e284f9c3bdb39909b354 Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.680597 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-config-data\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.680707 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.680749 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-scripts\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.680852 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nkr5\" (UniqueName: \"kubernetes.io/projected/d73ecea1-981c-4b0a-91f1-ace506d48483-kube-api-access-6nkr5\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.687801 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.688064 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-scripts\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.688746 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-config-data\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.700600 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nkr5\" (UniqueName: \"kubernetes.io/projected/d73ecea1-981c-4b0a-91f1-ace506d48483-kube-api-access-6nkr5\") pod \"nova-cell1-conductor-db-sync-2b5qb\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.824684 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-w87xf"] Dec 08 15:06:24 crc kubenswrapper[4894]: W1208 15:06:24.827539 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca71594_fdd4_4d9b_93b2_81f739476f9d.slice/crio-99c161f07e0d6b70312de02dfbb99d95506ecf02d13f19c81b9184d7aa99b9e9 WatchSource:0}: Error finding container 99c161f07e0d6b70312de02dfbb99d95506ecf02d13f19c81b9184d7aa99b9e9: Status 404 returned error can't find the container with id 99c161f07e0d6b70312de02dfbb99d95506ecf02d13f19c81b9184d7aa99b9e9 Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.834644 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:24 crc kubenswrapper[4894]: I1208 15:06:24.916258 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.184660 4894 generic.go:334] "Generic (PLEG): container finished" podID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerID="2bbf1942037b9d75778f63717d403ec066afc3594475ee2d3d4e143903706269" exitCode=0 Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.184833 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" event={"ID":"0c18cbf5-83bf-424d-9870-1f921b25a9e2","Type":"ContainerDied","Data":"2bbf1942037b9d75778f63717d403ec066afc3594475ee2d3d4e143903706269"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.185483 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" event={"ID":"0c18cbf5-83bf-424d-9870-1f921b25a9e2","Type":"ContainerStarted","Data":"078f6a58921091195e9969e36d66805ae99e3a606b577357d610ffcc74780a72"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.193021 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"612e0da4-aaaa-4050-be6d-ad14bae102be","Type":"ContainerStarted","Data":"e28407ed1fb3661f2d151147a973905b6a18fc09cf46e284f9c3bdb39909b354"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.226171 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mjdxz" event={"ID":"06425613-4821-4a3d-9507-f6176518464b","Type":"ContainerStarted","Data":"5c93a8fb060ccaf2f3a853961218545d48cece43a590aea4fc345486ed51c89f"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.226485 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mjdxz" event={"ID":"06425613-4821-4a3d-9507-f6176518464b","Type":"ContainerStarted","Data":"f3441805b723086babfdb0f7478f3fb15c548bb53fe568ea3caaf5746b5d65ac"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.226505 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca71594-fdd4-4d9b-93b2-81f739476f9d","Type":"ContainerStarted","Data":"99c161f07e0d6b70312de02dfbb99d95506ecf02d13f19c81b9184d7aa99b9e9"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.226520 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31","Type":"ContainerStarted","Data":"bc2b143f57fa7cbefbbd6b7dbae10b066c9e4d8c3efb44d2ce92d1d5714e6891"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.226532 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8228dc10-46b5-4855-a46c-aa6e16d09735","Type":"ContainerStarted","Data":"953f1d726d0501a0c3173a373e5b5e48f3e061823b0d3e22fe859857235a06ac"} Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.229970 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-mjdxz" podStartSLOduration=2.229951174 podStartE2EDuration="2.229951174s" podCreationTimestamp="2025-12-08 15:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:25.227733435 +0000 UTC m=+1206.327739550" watchObservedRunningTime="2025-12-08 15:06:25.229951174 +0000 UTC m=+1206.329957289" Dec 08 15:06:25 crc kubenswrapper[4894]: I1208 15:06:25.387688 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2b5qb"] Dec 08 15:06:25 crc kubenswrapper[4894]: W1208 15:06:25.409000 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd73ecea1_981c_4b0a_91f1_ace506d48483.slice/crio-5536027a6f7cbca181cce18c096e12f845fc823e42120183637aaee3ad9f079b WatchSource:0}: Error finding container 5536027a6f7cbca181cce18c096e12f845fc823e42120183637aaee3ad9f079b: Status 404 returned error can't find the container with id 5536027a6f7cbca181cce18c096e12f845fc823e42120183637aaee3ad9f079b Dec 08 15:06:26 crc kubenswrapper[4894]: I1208 15:06:26.244837 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" event={"ID":"0c18cbf5-83bf-424d-9870-1f921b25a9e2","Type":"ContainerStarted","Data":"9a06884443733ed8538c8804af17f12438fa67ec5fae8fc3df0592d909115676"} Dec 08 15:06:26 crc kubenswrapper[4894]: I1208 15:06:26.246160 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:26 crc kubenswrapper[4894]: I1208 15:06:26.250422 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" event={"ID":"d73ecea1-981c-4b0a-91f1-ace506d48483","Type":"ContainerStarted","Data":"6348e27b10880292ffc95b16fe26fc6300ab2e14fa8daa23a5b55a13e3a76e2b"} Dec 08 15:06:26 crc kubenswrapper[4894]: I1208 15:06:26.250453 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" event={"ID":"d73ecea1-981c-4b0a-91f1-ace506d48483","Type":"ContainerStarted","Data":"5536027a6f7cbca181cce18c096e12f845fc823e42120183637aaee3ad9f079b"} Dec 08 15:06:26 crc kubenswrapper[4894]: I1208 15:06:26.274253 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" podStartSLOduration=3.27423355 podStartE2EDuration="3.27423355s" podCreationTimestamp="2025-12-08 15:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:26.267175719 +0000 UTC m=+1207.367181824" watchObservedRunningTime="2025-12-08 15:06:26.27423355 +0000 UTC m=+1207.374239665" Dec 08 15:06:26 crc kubenswrapper[4894]: I1208 15:06:26.306310 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" podStartSLOduration=2.3062904140000002 podStartE2EDuration="2.306290414s" podCreationTimestamp="2025-12-08 15:06:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:26.288665512 +0000 UTC m=+1207.388671637" watchObservedRunningTime="2025-12-08 15:06:26.306290414 +0000 UTC m=+1207.406296529" Dec 08 15:06:27 crc kubenswrapper[4894]: I1208 15:06:27.999202 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:28 crc kubenswrapper[4894]: I1208 15:06:28.059328 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.309471 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-log" containerID="cri-o://9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480" gracePeriod=30 Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.309350 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31","Type":"ContainerStarted","Data":"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038"} Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.311094 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31","Type":"ContainerStarted","Data":"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480"} Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.310744 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-metadata" containerID="cri-o://28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038" gracePeriod=30 Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.315590 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8228dc10-46b5-4855-a46c-aa6e16d09735","Type":"ContainerStarted","Data":"fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8"} Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.315731 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8228dc10-46b5-4855-a46c-aa6e16d09735" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8" gracePeriod=30 Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.321838 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"612e0da4-aaaa-4050-be6d-ad14bae102be","Type":"ContainerStarted","Data":"252160400ed2b8328b7d9cf5aa1a27f609ca80f34770259b462f901fbd17e9fa"} Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.321881 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"612e0da4-aaaa-4050-be6d-ad14bae102be","Type":"ContainerStarted","Data":"1044c08a8de9fc242c3ff230045c77b9f2828ed456ba450a77b33fbc912a13a6"} Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.324062 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca71594-fdd4-4d9b-93b2-81f739476f9d","Type":"ContainerStarted","Data":"61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d"} Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.337254 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.57116685 podStartE2EDuration="6.337237129s" podCreationTimestamp="2025-12-08 15:06:23 +0000 UTC" firstStartedPulling="2025-12-08 15:06:24.341478698 +0000 UTC m=+1205.441484823" lastFinishedPulling="2025-12-08 15:06:28.107548987 +0000 UTC m=+1209.207555102" observedRunningTime="2025-12-08 15:06:29.332638596 +0000 UTC m=+1210.432644711" watchObservedRunningTime="2025-12-08 15:06:29.337237129 +0000 UTC m=+1210.437243244" Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.354189 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.076880899 podStartE2EDuration="6.35416948s" podCreationTimestamp="2025-12-08 15:06:23 +0000 UTC" firstStartedPulling="2025-12-08 15:06:24.830723761 +0000 UTC m=+1205.930729876" lastFinishedPulling="2025-12-08 15:06:28.108012342 +0000 UTC m=+1209.208018457" observedRunningTime="2025-12-08 15:06:29.353265741 +0000 UTC m=+1210.453271856" watchObservedRunningTime="2025-12-08 15:06:29.35416948 +0000 UTC m=+1210.454175595" Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.381046 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.821936294 podStartE2EDuration="6.38102375s" podCreationTimestamp="2025-12-08 15:06:23 +0000 UTC" firstStartedPulling="2025-12-08 15:06:24.547929494 +0000 UTC m=+1205.647935609" lastFinishedPulling="2025-12-08 15:06:28.10701695 +0000 UTC m=+1209.207023065" observedRunningTime="2025-12-08 15:06:29.374165686 +0000 UTC m=+1210.474171801" watchObservedRunningTime="2025-12-08 15:06:29.38102375 +0000 UTC m=+1210.481029855" Dec 08 15:06:29 crc kubenswrapper[4894]: I1208 15:06:29.400984 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.93932798 podStartE2EDuration="6.400965605s" podCreationTimestamp="2025-12-08 15:06:23 +0000 UTC" firstStartedPulling="2025-12-08 15:06:24.649756963 +0000 UTC m=+1205.749763078" lastFinishedPulling="2025-12-08 15:06:28.111394588 +0000 UTC m=+1209.211400703" observedRunningTime="2025-12-08 15:06:29.392416737 +0000 UTC m=+1210.492422852" watchObservedRunningTime="2025-12-08 15:06:29.400965605 +0000 UTC m=+1210.500971720" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.345301 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.370607 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.495711 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-logs\") pod \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.495855 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-config-data\") pod \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.495889 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-combined-ca-bundle\") pod \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.496077 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftlp2\" (UniqueName: \"kubernetes.io/projected/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-kube-api-access-ftlp2\") pod \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\" (UID: \"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31\") " Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.499922 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-logs" (OuterVolumeSpecName: "logs") pod "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" (UID: "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.511266 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-kube-api-access-ftlp2" (OuterVolumeSpecName: "kube-api-access-ftlp2") pod "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" (UID: "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31"). InnerVolumeSpecName "kube-api-access-ftlp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.521170 4894 generic.go:334] "Generic (PLEG): container finished" podID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerID="28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038" exitCode=0 Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.521214 4894 generic.go:334] "Generic (PLEG): container finished" podID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerID="9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480" exitCode=143 Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.521241 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31","Type":"ContainerDied","Data":"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038"} Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.521277 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31","Type":"ContainerDied","Data":"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480"} Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.521294 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31","Type":"ContainerDied","Data":"bc2b143f57fa7cbefbbd6b7dbae10b066c9e4d8c3efb44d2ce92d1d5714e6891"} Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.521312 4894 scope.go:117] "RemoveContainer" containerID="28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.521500 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.535044 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-config-data" (OuterVolumeSpecName: "config-data") pod "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" (UID: "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.547853 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" (UID: "0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.597926 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.597966 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.597979 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.597991 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftlp2\" (UniqueName: \"kubernetes.io/projected/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31-kube-api-access-ftlp2\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.648838 4894 scope.go:117] "RemoveContainer" containerID="9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.674692 4894 scope.go:117] "RemoveContainer" containerID="28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038" Dec 08 15:06:31 crc kubenswrapper[4894]: E1208 15:06:31.675179 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038\": container with ID starting with 28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038 not found: ID does not exist" containerID="28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.675216 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038"} err="failed to get container status \"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038\": rpc error: code = NotFound desc = could not find container \"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038\": container with ID starting with 28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038 not found: ID does not exist" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.675246 4894 scope.go:117] "RemoveContainer" containerID="9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480" Dec 08 15:06:31 crc kubenswrapper[4894]: E1208 15:06:31.675601 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480\": container with ID starting with 9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480 not found: ID does not exist" containerID="9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.675626 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480"} err="failed to get container status \"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480\": rpc error: code = NotFound desc = could not find container \"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480\": container with ID starting with 9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480 not found: ID does not exist" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.675642 4894 scope.go:117] "RemoveContainer" containerID="28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.676099 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038"} err="failed to get container status \"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038\": rpc error: code = NotFound desc = could not find container \"28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038\": container with ID starting with 28cdc519e9b7685bf772649193d3092615ab8be4c74caea673e36bb0735eb038 not found: ID does not exist" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.676119 4894 scope.go:117] "RemoveContainer" containerID="9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.676412 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480"} err="failed to get container status \"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480\": rpc error: code = NotFound desc = could not find container \"9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480\": container with ID starting with 9da9ffae35f78c88b5d15184ddcb583c20f08f47490ad7d08c7e5be96d726480 not found: ID does not exist" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.861497 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.874305 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.897715 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:31 crc kubenswrapper[4894]: E1208 15:06:31.898196 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-metadata" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.898224 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-metadata" Dec 08 15:06:31 crc kubenswrapper[4894]: E1208 15:06:31.898260 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-log" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.898268 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-log" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.898471 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-metadata" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.898507 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" containerName="nova-metadata-log" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.899570 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.905177 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.905295 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 08 15:06:31 crc kubenswrapper[4894]: I1208 15:06:31.928079 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.005274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.005388 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.005415 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfbr9\" (UniqueName: \"kubernetes.io/projected/b0ce9208-2a41-43a7-92a8-f72e7c262b22-kube-api-access-sfbr9\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.005449 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ce9208-2a41-43a7-92a8-f72e7c262b22-logs\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.005503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-config-data\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.108002 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.108090 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.108110 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfbr9\" (UniqueName: \"kubernetes.io/projected/b0ce9208-2a41-43a7-92a8-f72e7c262b22-kube-api-access-sfbr9\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.108133 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ce9208-2a41-43a7-92a8-f72e7c262b22-logs\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.108178 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-config-data\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.109651 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ce9208-2a41-43a7-92a8-f72e7c262b22-logs\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.114458 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.115030 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-config-data\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.115120 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.127442 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfbr9\" (UniqueName: \"kubernetes.io/projected/b0ce9208-2a41-43a7-92a8-f72e7c262b22-kube-api-access-sfbr9\") pod \"nova-metadata-0\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.231513 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:32 crc kubenswrapper[4894]: I1208 15:06:32.718151 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:32 crc kubenswrapper[4894]: W1208 15:06:32.722849 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0ce9208_2a41_43a7_92a8_f72e7c262b22.slice/crio-51cae126fd45539b8cb2167135d1386278159be1417c2399954ed2f0b6419270 WatchSource:0}: Error finding container 51cae126fd45539b8cb2167135d1386278159be1417c2399954ed2f0b6419270: Status 404 returned error can't find the container with id 51cae126fd45539b8cb2167135d1386278159be1417c2399954ed2f0b6419270 Dec 08 15:06:33 crc kubenswrapper[4894]: I1208 15:06:33.210994 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31" path="/var/lib/kubelet/pods/0dcc0f5d-9f90-44ab-9bc9-f7deb5c8fa31/volumes" Dec 08 15:06:33 crc kubenswrapper[4894]: I1208 15:06:33.563572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0ce9208-2a41-43a7-92a8-f72e7c262b22","Type":"ContainerStarted","Data":"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880"} Dec 08 15:06:33 crc kubenswrapper[4894]: I1208 15:06:33.564042 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0ce9208-2a41-43a7-92a8-f72e7c262b22","Type":"ContainerStarted","Data":"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d"} Dec 08 15:06:33 crc kubenswrapper[4894]: I1208 15:06:33.564068 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0ce9208-2a41-43a7-92a8-f72e7c262b22","Type":"ContainerStarted","Data":"51cae126fd45539b8cb2167135d1386278159be1417c2399954ed2f0b6419270"} Dec 08 15:06:33 crc kubenswrapper[4894]: I1208 15:06:33.602626 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6026064460000002 podStartE2EDuration="2.602606446s" podCreationTimestamp="2025-12-08 15:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:33.59731546 +0000 UTC m=+1214.697321575" watchObservedRunningTime="2025-12-08 15:06:33.602606446 +0000 UTC m=+1214.702612561" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.000277 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.069072 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.069135 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.089988 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.118001 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.118063 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.160084 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2gpp5"] Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.160371 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" podUID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerName="dnsmasq-dns" containerID="cri-o://c56caa9679e2e7ffd69bb6faf7336d6210d35a05890dcd96203884cbdc6cc5bb" gracePeriod=10 Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.180028 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.579636 4894 generic.go:334] "Generic (PLEG): container finished" podID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerID="c56caa9679e2e7ffd69bb6faf7336d6210d35a05890dcd96203884cbdc6cc5bb" exitCode=0 Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.580056 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" event={"ID":"b6ea0e91-5015-4954-ac06-8dad5fd37190","Type":"ContainerDied","Data":"c56caa9679e2e7ffd69bb6faf7336d6210d35a05890dcd96203884cbdc6cc5bb"} Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.582676 4894 generic.go:334] "Generic (PLEG): container finished" podID="06425613-4821-4a3d-9507-f6176518464b" containerID="5c93a8fb060ccaf2f3a853961218545d48cece43a590aea4fc345486ed51c89f" exitCode=0 Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.583809 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mjdxz" event={"ID":"06425613-4821-4a3d-9507-f6176518464b","Type":"ContainerDied","Data":"5c93a8fb060ccaf2f3a853961218545d48cece43a590aea4fc345486ed51c89f"} Dec 08 15:06:34 crc kubenswrapper[4894]: I1208 15:06:34.643971 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.016390 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.153041 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.153087 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.177460 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-config\") pod \"b6ea0e91-5015-4954-ac06-8dad5fd37190\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.177594 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-nb\") pod \"b6ea0e91-5015-4954-ac06-8dad5fd37190\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.177662 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-sb\") pod \"b6ea0e91-5015-4954-ac06-8dad5fd37190\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.177684 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-swift-storage-0\") pod \"b6ea0e91-5015-4954-ac06-8dad5fd37190\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.177747 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-svc\") pod \"b6ea0e91-5015-4954-ac06-8dad5fd37190\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.177805 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmf4n\" (UniqueName: \"kubernetes.io/projected/b6ea0e91-5015-4954-ac06-8dad5fd37190-kube-api-access-lmf4n\") pod \"b6ea0e91-5015-4954-ac06-8dad5fd37190\" (UID: \"b6ea0e91-5015-4954-ac06-8dad5fd37190\") " Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.191111 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6ea0e91-5015-4954-ac06-8dad5fd37190-kube-api-access-lmf4n" (OuterVolumeSpecName: "kube-api-access-lmf4n") pod "b6ea0e91-5015-4954-ac06-8dad5fd37190" (UID: "b6ea0e91-5015-4954-ac06-8dad5fd37190"). InnerVolumeSpecName "kube-api-access-lmf4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.253448 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-config" (OuterVolumeSpecName: "config") pod "b6ea0e91-5015-4954-ac06-8dad5fd37190" (UID: "b6ea0e91-5015-4954-ac06-8dad5fd37190"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.270884 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b6ea0e91-5015-4954-ac06-8dad5fd37190" (UID: "b6ea0e91-5015-4954-ac06-8dad5fd37190"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.275860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b6ea0e91-5015-4954-ac06-8dad5fd37190" (UID: "b6ea0e91-5015-4954-ac06-8dad5fd37190"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.282137 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.282181 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.282192 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.282202 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmf4n\" (UniqueName: \"kubernetes.io/projected/b6ea0e91-5015-4954-ac06-8dad5fd37190-kube-api-access-lmf4n\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.291642 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b6ea0e91-5015-4954-ac06-8dad5fd37190" (UID: "b6ea0e91-5015-4954-ac06-8dad5fd37190"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.322352 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b6ea0e91-5015-4954-ac06-8dad5fd37190" (UID: "b6ea0e91-5015-4954-ac06-8dad5fd37190"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.384399 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.384436 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b6ea0e91-5015-4954-ac06-8dad5fd37190-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.599185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" event={"ID":"b6ea0e91-5015-4954-ac06-8dad5fd37190","Type":"ContainerDied","Data":"51b48f356c1cf26715b90d4d3ffcc43efca56e233c0809a116c7a59d9b8c371b"} Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.599222 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-2gpp5" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.599257 4894 scope.go:117] "RemoveContainer" containerID="c56caa9679e2e7ffd69bb6faf7336d6210d35a05890dcd96203884cbdc6cc5bb" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.687583 4894 scope.go:117] "RemoveContainer" containerID="bc20d2242126f668c23a26db8565128f10768432421c56abe86521cf8c071199" Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.700877 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2gpp5"] Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.771017 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-2gpp5"] Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.905506 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:06:35 crc kubenswrapper[4894]: I1208 15:06:35.907946 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="102cc0a0-d75c-4050-a100-6c97351084dd" containerName="kube-state-metrics" containerID="cri-o://2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff" gracePeriod=30 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.104430 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.198551 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-config-data\") pod \"06425613-4821-4a3d-9507-f6176518464b\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.198635 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-scripts\") pod \"06425613-4821-4a3d-9507-f6176518464b\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.198722 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-combined-ca-bundle\") pod \"06425613-4821-4a3d-9507-f6176518464b\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.198979 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wvqk\" (UniqueName: \"kubernetes.io/projected/06425613-4821-4a3d-9507-f6176518464b-kube-api-access-6wvqk\") pod \"06425613-4821-4a3d-9507-f6176518464b\" (UID: \"06425613-4821-4a3d-9507-f6176518464b\") " Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.204515 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06425613-4821-4a3d-9507-f6176518464b-kube-api-access-6wvqk" (OuterVolumeSpecName: "kube-api-access-6wvqk") pod "06425613-4821-4a3d-9507-f6176518464b" (UID: "06425613-4821-4a3d-9507-f6176518464b"). InnerVolumeSpecName "kube-api-access-6wvqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.223221 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-scripts" (OuterVolumeSpecName: "scripts") pod "06425613-4821-4a3d-9507-f6176518464b" (UID: "06425613-4821-4a3d-9507-f6176518464b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.262124 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-config-data" (OuterVolumeSpecName: "config-data") pod "06425613-4821-4a3d-9507-f6176518464b" (UID: "06425613-4821-4a3d-9507-f6176518464b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.290398 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06425613-4821-4a3d-9507-f6176518464b" (UID: "06425613-4821-4a3d-9507-f6176518464b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.303944 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.303972 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.303981 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06425613-4821-4a3d-9507-f6176518464b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.303993 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wvqk\" (UniqueName: \"kubernetes.io/projected/06425613-4821-4a3d-9507-f6176518464b-kube-api-access-6wvqk\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.402935 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.506852 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjnzc\" (UniqueName: \"kubernetes.io/projected/102cc0a0-d75c-4050-a100-6c97351084dd-kube-api-access-hjnzc\") pod \"102cc0a0-d75c-4050-a100-6c97351084dd\" (UID: \"102cc0a0-d75c-4050-a100-6c97351084dd\") " Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.523977 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/102cc0a0-d75c-4050-a100-6c97351084dd-kube-api-access-hjnzc" (OuterVolumeSpecName: "kube-api-access-hjnzc") pod "102cc0a0-d75c-4050-a100-6c97351084dd" (UID: "102cc0a0-d75c-4050-a100-6c97351084dd"). InnerVolumeSpecName "kube-api-access-hjnzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.609174 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjnzc\" (UniqueName: \"kubernetes.io/projected/102cc0a0-d75c-4050-a100-6c97351084dd-kube-api-access-hjnzc\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.611502 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mjdxz" event={"ID":"06425613-4821-4a3d-9507-f6176518464b","Type":"ContainerDied","Data":"f3441805b723086babfdb0f7478f3fb15c548bb53fe568ea3caaf5746b5d65ac"} Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.611541 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3441805b723086babfdb0f7478f3fb15c548bb53fe568ea3caaf5746b5d65ac" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.611548 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mjdxz" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.613306 4894 generic.go:334] "Generic (PLEG): container finished" podID="d73ecea1-981c-4b0a-91f1-ace506d48483" containerID="6348e27b10880292ffc95b16fe26fc6300ab2e14fa8daa23a5b55a13e3a76e2b" exitCode=0 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.613427 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" event={"ID":"d73ecea1-981c-4b0a-91f1-ace506d48483","Type":"ContainerDied","Data":"6348e27b10880292ffc95b16fe26fc6300ab2e14fa8daa23a5b55a13e3a76e2b"} Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.615335 4894 generic.go:334] "Generic (PLEG): container finished" podID="102cc0a0-d75c-4050-a100-6c97351084dd" containerID="2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff" exitCode=2 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.615361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"102cc0a0-d75c-4050-a100-6c97351084dd","Type":"ContainerDied","Data":"2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff"} Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.615379 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"102cc0a0-d75c-4050-a100-6c97351084dd","Type":"ContainerDied","Data":"b5c06232fa14fdfb018353f51c59e8ebf8256a7ed50e052b8cdfe2c1291ab716"} Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.615397 4894 scope.go:117] "RemoveContainer" containerID="2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.615408 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.665595 4894 scope.go:117] "RemoveContainer" containerID="2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff" Dec 08 15:06:36 crc kubenswrapper[4894]: E1208 15:06:36.666325 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff\": container with ID starting with 2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff not found: ID does not exist" containerID="2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.666430 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff"} err="failed to get container status \"2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff\": rpc error: code = NotFound desc = could not find container \"2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff\": container with ID starting with 2b0629675105e2b6f0cf642087ac375ecb488809aeb165e151b6568ab71d3cff not found: ID does not exist" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.673930 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.692714 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704217 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:06:36 crc kubenswrapper[4894]: E1208 15:06:36.704589 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerName="init" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704603 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerName="init" Dec 08 15:06:36 crc kubenswrapper[4894]: E1208 15:06:36.704622 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerName="dnsmasq-dns" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704628 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerName="dnsmasq-dns" Dec 08 15:06:36 crc kubenswrapper[4894]: E1208 15:06:36.704659 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06425613-4821-4a3d-9507-f6176518464b" containerName="nova-manage" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704665 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="06425613-4821-4a3d-9507-f6176518464b" containerName="nova-manage" Dec 08 15:06:36 crc kubenswrapper[4894]: E1208 15:06:36.704675 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102cc0a0-d75c-4050-a100-6c97351084dd" containerName="kube-state-metrics" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704681 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="102cc0a0-d75c-4050-a100-6c97351084dd" containerName="kube-state-metrics" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704866 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="06425613-4821-4a3d-9507-f6176518464b" containerName="nova-manage" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704880 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6ea0e91-5015-4954-ac06-8dad5fd37190" containerName="dnsmasq-dns" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.704899 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="102cc0a0-d75c-4050-a100-6c97351084dd" containerName="kube-state-metrics" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.705496 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.712336 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.712381 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.714840 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.815652 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.815836 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.815891 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.815947 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg88d\" (UniqueName: \"kubernetes.io/projected/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-api-access-lg88d\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.880327 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.880609 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-log" containerID="cri-o://1044c08a8de9fc242c3ff230045c77b9f2828ed456ba450a77b33fbc912a13a6" gracePeriod=30 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.881123 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-api" containerID="cri-o://252160400ed2b8328b7d9cf5aa1a27f609ca80f34770259b462f901fbd17e9fa" gracePeriod=30 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.892040 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.892273 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3ca71594-fdd4-4d9b-93b2-81f739476f9d" containerName="nova-scheduler-scheduler" containerID="cri-o://61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" gracePeriod=30 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.905489 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.905711 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-log" containerID="cri-o://5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d" gracePeriod=30 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.905857 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-metadata" containerID="cri-o://85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880" gracePeriod=30 Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.919170 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.919275 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.919341 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg88d\" (UniqueName: \"kubernetes.io/projected/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-api-access-lg88d\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.919414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.923302 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.924490 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.933585 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:36 crc kubenswrapper[4894]: I1208 15:06:36.941785 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg88d\" (UniqueName: \"kubernetes.io/projected/5ea146ea-5684-46f7-850c-77df5b66c07b-kube-api-access-lg88d\") pod \"kube-state-metrics-0\" (UID: \"5ea146ea-5684-46f7-850c-77df5b66c07b\") " pod="openstack/kube-state-metrics-0" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.042410 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.213223 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="102cc0a0-d75c-4050-a100-6c97351084dd" path="/var/lib/kubelet/pods/102cc0a0-d75c-4050-a100-6c97351084dd/volumes" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.214168 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6ea0e91-5015-4954-ac06-8dad5fd37190" path="/var/lib/kubelet/pods/b6ea0e91-5015-4954-ac06-8dad5fd37190/volumes" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.233197 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.233241 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.547373 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 08 15:06:37 crc kubenswrapper[4894]: W1208 15:06:37.564551 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ea146ea_5684_46f7_850c_77df5b66c07b.slice/crio-86bde49f809edee5bf2596f9c58fcea0dc5fdb44012b2cf538ec82db3b314ae1 WatchSource:0}: Error finding container 86bde49f809edee5bf2596f9c58fcea0dc5fdb44012b2cf538ec82db3b314ae1: Status 404 returned error can't find the container with id 86bde49f809edee5bf2596f9c58fcea0dc5fdb44012b2cf538ec82db3b314ae1 Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.634085 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5ea146ea-5684-46f7-850c-77df5b66c07b","Type":"ContainerStarted","Data":"86bde49f809edee5bf2596f9c58fcea0dc5fdb44012b2cf538ec82db3b314ae1"} Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.636665 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.641965 4894 generic.go:334] "Generic (PLEG): container finished" podID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerID="1044c08a8de9fc242c3ff230045c77b9f2828ed456ba450a77b33fbc912a13a6" exitCode=143 Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.642058 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"612e0da4-aaaa-4050-be6d-ad14bae102be","Type":"ContainerDied","Data":"1044c08a8de9fc242c3ff230045c77b9f2828ed456ba450a77b33fbc912a13a6"} Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.648507 4894 generic.go:334] "Generic (PLEG): container finished" podID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerID="85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880" exitCode=0 Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.648569 4894 generic.go:334] "Generic (PLEG): container finished" podID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerID="5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d" exitCode=143 Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.648591 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0ce9208-2a41-43a7-92a8-f72e7c262b22","Type":"ContainerDied","Data":"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880"} Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.648702 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0ce9208-2a41-43a7-92a8-f72e7c262b22","Type":"ContainerDied","Data":"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d"} Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.648725 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b0ce9208-2a41-43a7-92a8-f72e7c262b22","Type":"ContainerDied","Data":"51cae126fd45539b8cb2167135d1386278159be1417c2399954ed2f0b6419270"} Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.648745 4894 scope.go:117] "RemoveContainer" containerID="85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.648627 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.702591 4894 scope.go:117] "RemoveContainer" containerID="5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.735489 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfbr9\" (UniqueName: \"kubernetes.io/projected/b0ce9208-2a41-43a7-92a8-f72e7c262b22-kube-api-access-sfbr9\") pod \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.735732 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-combined-ca-bundle\") pod \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.735891 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-nova-metadata-tls-certs\") pod \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.735931 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-config-data\") pod \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.735973 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ce9208-2a41-43a7-92a8-f72e7c262b22-logs\") pod \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\" (UID: \"b0ce9208-2a41-43a7-92a8-f72e7c262b22\") " Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.736883 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ce9208-2a41-43a7-92a8-f72e7c262b22-logs" (OuterVolumeSpecName: "logs") pod "b0ce9208-2a41-43a7-92a8-f72e7c262b22" (UID: "b0ce9208-2a41-43a7-92a8-f72e7c262b22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.753325 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ce9208-2a41-43a7-92a8-f72e7c262b22-kube-api-access-sfbr9" (OuterVolumeSpecName: "kube-api-access-sfbr9") pod "b0ce9208-2a41-43a7-92a8-f72e7c262b22" (UID: "b0ce9208-2a41-43a7-92a8-f72e7c262b22"). InnerVolumeSpecName "kube-api-access-sfbr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.778943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0ce9208-2a41-43a7-92a8-f72e7c262b22" (UID: "b0ce9208-2a41-43a7-92a8-f72e7c262b22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.796056 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-config-data" (OuterVolumeSpecName: "config-data") pod "b0ce9208-2a41-43a7-92a8-f72e7c262b22" (UID: "b0ce9208-2a41-43a7-92a8-f72e7c262b22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.839992 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.840039 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0ce9208-2a41-43a7-92a8-f72e7c262b22-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.840058 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfbr9\" (UniqueName: \"kubernetes.io/projected/b0ce9208-2a41-43a7-92a8-f72e7c262b22-kube-api-access-sfbr9\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.840068 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.857156 4894 scope.go:117] "RemoveContainer" containerID="85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880" Dec 08 15:06:37 crc kubenswrapper[4894]: E1208 15:06:37.864246 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880\": container with ID starting with 85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880 not found: ID does not exist" containerID="85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.864299 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880"} err="failed to get container status \"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880\": rpc error: code = NotFound desc = could not find container \"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880\": container with ID starting with 85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880 not found: ID does not exist" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.864332 4894 scope.go:117] "RemoveContainer" containerID="5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d" Dec 08 15:06:37 crc kubenswrapper[4894]: E1208 15:06:37.869238 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d\": container with ID starting with 5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d not found: ID does not exist" containerID="5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.869302 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d"} err="failed to get container status \"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d\": rpc error: code = NotFound desc = could not find container \"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d\": container with ID starting with 5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d not found: ID does not exist" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.869338 4894 scope.go:117] "RemoveContainer" containerID="85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.869625 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880"} err="failed to get container status \"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880\": rpc error: code = NotFound desc = could not find container \"85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880\": container with ID starting with 85c8f7e84de8cd61d613c2b91f4cc8ae4a17501417feb3a7a69b10fa38961880 not found: ID does not exist" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.869645 4894 scope.go:117] "RemoveContainer" containerID="5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.869830 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d"} err="failed to get container status \"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d\": rpc error: code = NotFound desc = could not find container \"5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d\": container with ID starting with 5483e285758a5d4f6a9235667d8a26327c713f5c8901800d7cb9ca277d9a491d not found: ID does not exist" Dec 08 15:06:37 crc kubenswrapper[4894]: I1208 15:06:37.971954 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b0ce9208-2a41-43a7-92a8-f72e7c262b22" (UID: "b0ce9208-2a41-43a7-92a8-f72e7c262b22"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.043412 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0ce9208-2a41-43a7-92a8-f72e7c262b22-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.112066 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.145671 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-combined-ca-bundle\") pod \"d73ecea1-981c-4b0a-91f1-ace506d48483\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.145764 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nkr5\" (UniqueName: \"kubernetes.io/projected/d73ecea1-981c-4b0a-91f1-ace506d48483-kube-api-access-6nkr5\") pod \"d73ecea1-981c-4b0a-91f1-ace506d48483\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.145806 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-scripts\") pod \"d73ecea1-981c-4b0a-91f1-ace506d48483\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.145889 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-config-data\") pod \"d73ecea1-981c-4b0a-91f1-ace506d48483\" (UID: \"d73ecea1-981c-4b0a-91f1-ace506d48483\") " Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.150174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73ecea1-981c-4b0a-91f1-ace506d48483-kube-api-access-6nkr5" (OuterVolumeSpecName: "kube-api-access-6nkr5") pod "d73ecea1-981c-4b0a-91f1-ace506d48483" (UID: "d73ecea1-981c-4b0a-91f1-ace506d48483"). InnerVolumeSpecName "kube-api-access-6nkr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.158221 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-scripts" (OuterVolumeSpecName: "scripts") pod "d73ecea1-981c-4b0a-91f1-ace506d48483" (UID: "d73ecea1-981c-4b0a-91f1-ace506d48483"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.191769 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-config-data" (OuterVolumeSpecName: "config-data") pod "d73ecea1-981c-4b0a-91f1-ace506d48483" (UID: "d73ecea1-981c-4b0a-91f1-ace506d48483"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.202194 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d73ecea1-981c-4b0a-91f1-ace506d48483" (UID: "d73ecea1-981c-4b0a-91f1-ace506d48483"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.252789 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.252855 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nkr5\" (UniqueName: \"kubernetes.io/projected/d73ecea1-981c-4b0a-91f1-ace506d48483-kube-api-access-6nkr5\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.252891 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.252965 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d73ecea1-981c-4b0a-91f1-ace506d48483-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.337642 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.379550 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.407667 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:38 crc kubenswrapper[4894]: E1208 15:06:38.408137 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73ecea1-981c-4b0a-91f1-ace506d48483" containerName="nova-cell1-conductor-db-sync" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.408161 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73ecea1-981c-4b0a-91f1-ace506d48483" containerName="nova-cell1-conductor-db-sync" Dec 08 15:06:38 crc kubenswrapper[4894]: E1208 15:06:38.408189 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-metadata" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.408200 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-metadata" Dec 08 15:06:38 crc kubenswrapper[4894]: E1208 15:06:38.408238 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-log" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.408247 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-log" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.408459 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d73ecea1-981c-4b0a-91f1-ace506d48483" containerName="nova-cell1-conductor-db-sync" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.408486 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-metadata" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.408497 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" containerName="nova-metadata-log" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.409723 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.412366 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.412771 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.418651 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.460339 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wztfv\" (UniqueName: \"kubernetes.io/projected/c42e42b6-4389-4938-973d-62cfc12cb948-kube-api-access-wztfv\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.460388 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.460424 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-config-data\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.460539 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.460640 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42e42b6-4389-4938-973d-62cfc12cb948-logs\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.562561 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wztfv\" (UniqueName: \"kubernetes.io/projected/c42e42b6-4389-4938-973d-62cfc12cb948-kube-api-access-wztfv\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.562619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.562664 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-config-data\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.562692 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.562743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42e42b6-4389-4938-973d-62cfc12cb948-logs\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.563399 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42e42b6-4389-4938-973d-62cfc12cb948-logs\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.567889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-config-data\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.568665 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.578626 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.578742 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wztfv\" (UniqueName: \"kubernetes.io/projected/c42e42b6-4389-4938-973d-62cfc12cb948-kube-api-access-wztfv\") pod \"nova-metadata-0\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.664568 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5ea146ea-5684-46f7-850c-77df5b66c07b","Type":"ContainerStarted","Data":"d4d7cd4e558f4b62d385f80e6ee77bf2eb174d84751731e9634bda7c43a1ee9a"} Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.664892 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.666395 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" event={"ID":"d73ecea1-981c-4b0a-91f1-ace506d48483","Type":"ContainerDied","Data":"5536027a6f7cbca181cce18c096e12f845fc823e42120183637aaee3ad9f079b"} Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.666421 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5536027a6f7cbca181cce18c096e12f845fc823e42120183637aaee3ad9f079b" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.666469 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2b5qb" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.692143 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.174673897 podStartE2EDuration="2.692126823s" podCreationTimestamp="2025-12-08 15:06:36 +0000 UTC" firstStartedPulling="2025-12-08 15:06:37.582386168 +0000 UTC m=+1218.682392283" lastFinishedPulling="2025-12-08 15:06:38.099839094 +0000 UTC m=+1219.199845209" observedRunningTime="2025-12-08 15:06:38.689256513 +0000 UTC m=+1219.789262628" watchObservedRunningTime="2025-12-08 15:06:38.692126823 +0000 UTC m=+1219.792132938" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.716275 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.717518 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.720098 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.730219 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.739731 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.767233 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhkf4\" (UniqueName: \"kubernetes.io/projected/e939bafe-db68-41d7-8505-651b08bd1ffd-kube-api-access-jhkf4\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.767329 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e939bafe-db68-41d7-8505-651b08bd1ffd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.767400 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e939bafe-db68-41d7-8505-651b08bd1ffd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.794585 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.795518 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-central-agent" containerID="cri-o://1efe09e6136fe6cd9c2fa9145b6e5de05a578f944d900052be192e81e2e268f4" gracePeriod=30 Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.795867 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="proxy-httpd" containerID="cri-o://2d8a3b0738fa1d5e7b6049581bbb6bac10282d01d55139b385b2bc0b59ec069e" gracePeriod=30 Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.795925 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-notification-agent" containerID="cri-o://4ffcb97f784d42b5dab0f78f0d8f43a46304afac4c73a3419b869d2dc795f47f" gracePeriod=30 Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.795992 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="sg-core" containerID="cri-o://e04d41dddd19aca575f13d7970b6b84068d355c9542d66c26cf40079d86c3c57" gracePeriod=30 Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.868680 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhkf4\" (UniqueName: \"kubernetes.io/projected/e939bafe-db68-41d7-8505-651b08bd1ffd-kube-api-access-jhkf4\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.868739 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e939bafe-db68-41d7-8505-651b08bd1ffd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.868773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e939bafe-db68-41d7-8505-651b08bd1ffd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.876074 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e939bafe-db68-41d7-8505-651b08bd1ffd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.876325 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e939bafe-db68-41d7-8505-651b08bd1ffd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:38 crc kubenswrapper[4894]: I1208 15:06:38.890761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhkf4\" (UniqueName: \"kubernetes.io/projected/e939bafe-db68-41d7-8505-651b08bd1ffd-kube-api-access-jhkf4\") pod \"nova-cell1-conductor-0\" (UID: \"e939bafe-db68-41d7-8505-651b08bd1ffd\") " pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.036131 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:39 crc kubenswrapper[4894]: E1208 15:06:39.118427 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 15:06:39 crc kubenswrapper[4894]: E1208 15:06:39.119984 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 15:06:39 crc kubenswrapper[4894]: E1208 15:06:39.123557 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 08 15:06:39 crc kubenswrapper[4894]: E1208 15:06:39.123671 4894 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3ca71594-fdd4-4d9b-93b2-81f739476f9d" containerName="nova-scheduler-scheduler" Dec 08 15:06:39 crc kubenswrapper[4894]: E1208 15:06:39.132200 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081d95e0_0092_43e3_9517_43a8a5b62d52.slice/crio-conmon-e04d41dddd19aca575f13d7970b6b84068d355c9542d66c26cf40079d86c3c57.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081d95e0_0092_43e3_9517_43a8a5b62d52.slice/crio-e04d41dddd19aca575f13d7970b6b84068d355c9542d66c26cf40079d86c3c57.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081d95e0_0092_43e3_9517_43a8a5b62d52.slice/crio-conmon-2d8a3b0738fa1d5e7b6049581bbb6bac10282d01d55139b385b2bc0b59ec069e.scope\": RecentStats: unable to find data in memory cache]" Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.218059 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ce9208-2a41-43a7-92a8-f72e7c262b22" path="/var/lib/kubelet/pods/b0ce9208-2a41-43a7-92a8-f72e7c262b22/volumes" Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.309784 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.565292 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.692530 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e939bafe-db68-41d7-8505-651b08bd1ffd","Type":"ContainerStarted","Data":"0ba9d1da7d3150676aa85d9d806a1ce8b8ce7941a585ae25f3dffd94b41f719e"} Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701139 4894 generic.go:334] "Generic (PLEG): container finished" podID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerID="2d8a3b0738fa1d5e7b6049581bbb6bac10282d01d55139b385b2bc0b59ec069e" exitCode=0 Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701180 4894 generic.go:334] "Generic (PLEG): container finished" podID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerID="e04d41dddd19aca575f13d7970b6b84068d355c9542d66c26cf40079d86c3c57" exitCode=2 Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701190 4894 generic.go:334] "Generic (PLEG): container finished" podID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerID="4ffcb97f784d42b5dab0f78f0d8f43a46304afac4c73a3419b869d2dc795f47f" exitCode=0 Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701199 4894 generic.go:334] "Generic (PLEG): container finished" podID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerID="1efe09e6136fe6cd9c2fa9145b6e5de05a578f944d900052be192e81e2e268f4" exitCode=0 Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701284 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerDied","Data":"2d8a3b0738fa1d5e7b6049581bbb6bac10282d01d55139b385b2bc0b59ec069e"} Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701313 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerDied","Data":"e04d41dddd19aca575f13d7970b6b84068d355c9542d66c26cf40079d86c3c57"} Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701324 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerDied","Data":"4ffcb97f784d42b5dab0f78f0d8f43a46304afac4c73a3419b869d2dc795f47f"} Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.701333 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerDied","Data":"1efe09e6136fe6cd9c2fa9145b6e5de05a578f944d900052be192e81e2e268f4"} Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.707051 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c42e42b6-4389-4938-973d-62cfc12cb948","Type":"ContainerStarted","Data":"bcbdcae84484e1062a65610577466a82e188ab55618599224acdca8f3fca2fcf"} Dec 08 15:06:39 crc kubenswrapper[4894]: I1208 15:06:39.894686 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.014479 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-log-httpd\") pod \"081d95e0-0092-43e3-9517-43a8a5b62d52\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.014565 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcpcf\" (UniqueName: \"kubernetes.io/projected/081d95e0-0092-43e3-9517-43a8a5b62d52-kube-api-access-gcpcf\") pod \"081d95e0-0092-43e3-9517-43a8a5b62d52\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.014766 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-combined-ca-bundle\") pod \"081d95e0-0092-43e3-9517-43a8a5b62d52\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.014854 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-config-data\") pod \"081d95e0-0092-43e3-9517-43a8a5b62d52\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.014900 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-run-httpd\") pod \"081d95e0-0092-43e3-9517-43a8a5b62d52\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.014919 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-scripts\") pod \"081d95e0-0092-43e3-9517-43a8a5b62d52\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.014951 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-sg-core-conf-yaml\") pod \"081d95e0-0092-43e3-9517-43a8a5b62d52\" (UID: \"081d95e0-0092-43e3-9517-43a8a5b62d52\") " Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.015054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "081d95e0-0092-43e3-9517-43a8a5b62d52" (UID: "081d95e0-0092-43e3-9517-43a8a5b62d52"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.015347 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.016397 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "081d95e0-0092-43e3-9517-43a8a5b62d52" (UID: "081d95e0-0092-43e3-9517-43a8a5b62d52"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.020183 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081d95e0-0092-43e3-9517-43a8a5b62d52-kube-api-access-gcpcf" (OuterVolumeSpecName: "kube-api-access-gcpcf") pod "081d95e0-0092-43e3-9517-43a8a5b62d52" (UID: "081d95e0-0092-43e3-9517-43a8a5b62d52"). InnerVolumeSpecName "kube-api-access-gcpcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.022496 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-scripts" (OuterVolumeSpecName: "scripts") pod "081d95e0-0092-43e3-9517-43a8a5b62d52" (UID: "081d95e0-0092-43e3-9517-43a8a5b62d52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.071231 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "081d95e0-0092-43e3-9517-43a8a5b62d52" (UID: "081d95e0-0092-43e3-9517-43a8a5b62d52"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.117504 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/081d95e0-0092-43e3-9517-43a8a5b62d52-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.117544 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.117554 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.117573 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcpcf\" (UniqueName: \"kubernetes.io/projected/081d95e0-0092-43e3-9517-43a8a5b62d52-kube-api-access-gcpcf\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.161855 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-config-data" (OuterVolumeSpecName: "config-data") pod "081d95e0-0092-43e3-9517-43a8a5b62d52" (UID: "081d95e0-0092-43e3-9517-43a8a5b62d52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.178090 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "081d95e0-0092-43e3-9517-43a8a5b62d52" (UID: "081d95e0-0092-43e3-9517-43a8a5b62d52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.219516 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.219563 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081d95e0-0092-43e3-9517-43a8a5b62d52-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.718871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e939bafe-db68-41d7-8505-651b08bd1ffd","Type":"ContainerStarted","Data":"0d50de06a625e195f88c9bd9a66a86a11e9b4fb791aaba633f1dde5b7e65638f"} Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.718984 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.724013 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"081d95e0-0092-43e3-9517-43a8a5b62d52","Type":"ContainerDied","Data":"d1662c378fb52cdbe5b191f96a5c91dae438cfb61b0cc9251b704207a00ed34d"} Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.724063 4894 scope.go:117] "RemoveContainer" containerID="2d8a3b0738fa1d5e7b6049581bbb6bac10282d01d55139b385b2bc0b59ec069e" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.724068 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.734388 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c42e42b6-4389-4938-973d-62cfc12cb948","Type":"ContainerStarted","Data":"0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b"} Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.734431 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c42e42b6-4389-4938-973d-62cfc12cb948","Type":"ContainerStarted","Data":"d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa"} Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.749849 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.749831358 podStartE2EDuration="2.749831358s" podCreationTimestamp="2025-12-08 15:06:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:40.739694091 +0000 UTC m=+1221.839700206" watchObservedRunningTime="2025-12-08 15:06:40.749831358 +0000 UTC m=+1221.849837473" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.754767 4894 scope.go:117] "RemoveContainer" containerID="e04d41dddd19aca575f13d7970b6b84068d355c9542d66c26cf40079d86c3c57" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.763251 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.763224918 podStartE2EDuration="2.763224918s" podCreationTimestamp="2025-12-08 15:06:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:40.759106259 +0000 UTC m=+1221.859112374" watchObservedRunningTime="2025-12-08 15:06:40.763224918 +0000 UTC m=+1221.863231043" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.788376 4894 scope.go:117] "RemoveContainer" containerID="4ffcb97f784d42b5dab0f78f0d8f43a46304afac4c73a3419b869d2dc795f47f" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.791601 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.816382 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.822273 4894 scope.go:117] "RemoveContainer" containerID="1efe09e6136fe6cd9c2fa9145b6e5de05a578f944d900052be192e81e2e268f4" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833157 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:40 crc kubenswrapper[4894]: E1208 15:06:40.833594 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-central-agent" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833612 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-central-agent" Dec 08 15:06:40 crc kubenswrapper[4894]: E1208 15:06:40.833648 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="sg-core" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833655 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="sg-core" Dec 08 15:06:40 crc kubenswrapper[4894]: E1208 15:06:40.833668 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-notification-agent" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833675 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-notification-agent" Dec 08 15:06:40 crc kubenswrapper[4894]: E1208 15:06:40.833686 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="proxy-httpd" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833692 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="proxy-httpd" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833944 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-notification-agent" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833964 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="ceilometer-central-agent" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833978 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="sg-core" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.833988 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" containerName="proxy-httpd" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.835707 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.840587 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.840782 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.840985 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.843994 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850046 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850276 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-scripts\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850304 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-log-httpd\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850331 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850400 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-config-data\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850461 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850493 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-run-httpd\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.850533 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8mg6\" (UniqueName: \"kubernetes.io/projected/8c25b112-8906-4404-a824-e20084014c1b-kube-api-access-r8mg6\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-scripts\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952844 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-log-httpd\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952861 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952890 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-config-data\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952923 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952951 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-run-httpd\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.952974 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8mg6\" (UniqueName: \"kubernetes.io/projected/8c25b112-8906-4404-a824-e20084014c1b-kube-api-access-r8mg6\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.953920 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-run-httpd\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.953998 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-log-httpd\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.960609 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.961207 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-scripts\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.976644 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.986425 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-config-data\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.986544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:40 crc kubenswrapper[4894]: I1208 15:06:40.993477 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8mg6\" (UniqueName: \"kubernetes.io/projected/8c25b112-8906-4404-a824-e20084014c1b-kube-api-access-r8mg6\") pod \"ceilometer-0\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " pod="openstack/ceilometer-0" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.176568 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.213109 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081d95e0-0092-43e3-9517-43a8a5b62d52" path="/var/lib/kubelet/pods/081d95e0-0092-43e3-9517-43a8a5b62d52/volumes" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.606214 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.703945 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-config-data\") pod \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.704100 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-combined-ca-bundle\") pod \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.704281 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25csl\" (UniqueName: \"kubernetes.io/projected/3ca71594-fdd4-4d9b-93b2-81f739476f9d-kube-api-access-25csl\") pod \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\" (UID: \"3ca71594-fdd4-4d9b-93b2-81f739476f9d\") " Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.734153 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca71594-fdd4-4d9b-93b2-81f739476f9d-kube-api-access-25csl" (OuterVolumeSpecName: "kube-api-access-25csl") pod "3ca71594-fdd4-4d9b-93b2-81f739476f9d" (UID: "3ca71594-fdd4-4d9b-93b2-81f739476f9d"). InnerVolumeSpecName "kube-api-access-25csl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.787492 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.789793 4894 generic.go:334] "Generic (PLEG): container finished" podID="3ca71594-fdd4-4d9b-93b2-81f739476f9d" containerID="61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" exitCode=0 Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.789863 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca71594-fdd4-4d9b-93b2-81f739476f9d","Type":"ContainerDied","Data":"61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d"} Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.789886 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3ca71594-fdd4-4d9b-93b2-81f739476f9d","Type":"ContainerDied","Data":"99c161f07e0d6b70312de02dfbb99d95506ecf02d13f19c81b9184d7aa99b9e9"} Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.789906 4894 scope.go:117] "RemoveContainer" containerID="61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.789986 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.795498 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ca71594-fdd4-4d9b-93b2-81f739476f9d" (UID: "3ca71594-fdd4-4d9b-93b2-81f739476f9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.807885 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.807918 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25csl\" (UniqueName: \"kubernetes.io/projected/3ca71594-fdd4-4d9b-93b2-81f739476f9d-kube-api-access-25csl\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.811768 4894 generic.go:334] "Generic (PLEG): container finished" podID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerID="252160400ed2b8328b7d9cf5aa1a27f609ca80f34770259b462f901fbd17e9fa" exitCode=0 Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.811798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"612e0da4-aaaa-4050-be6d-ad14bae102be","Type":"ContainerDied","Data":"252160400ed2b8328b7d9cf5aa1a27f609ca80f34770259b462f901fbd17e9fa"} Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.835082 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-config-data" (OuterVolumeSpecName: "config-data") pod "3ca71594-fdd4-4d9b-93b2-81f739476f9d" (UID: "3ca71594-fdd4-4d9b-93b2-81f739476f9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.851332 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.857327 4894 scope.go:117] "RemoveContainer" containerID="61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" Dec 08 15:06:41 crc kubenswrapper[4894]: E1208 15:06:41.857929 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d\": container with ID starting with 61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d not found: ID does not exist" containerID="61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.857994 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d"} err="failed to get container status \"61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d\": rpc error: code = NotFound desc = could not find container \"61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d\": container with ID starting with 61cd647785c0fe5be37cc94b9b238b124aa8aa2e6cbea2243540b8fd3282591d not found: ID does not exist" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.909366 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-combined-ca-bundle\") pod \"612e0da4-aaaa-4050-be6d-ad14bae102be\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.909550 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data\") pod \"612e0da4-aaaa-4050-be6d-ad14bae102be\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.909589 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/612e0da4-aaaa-4050-be6d-ad14bae102be-logs\") pod \"612e0da4-aaaa-4050-be6d-ad14bae102be\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.909613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86r6v\" (UniqueName: \"kubernetes.io/projected/612e0da4-aaaa-4050-be6d-ad14bae102be-kube-api-access-86r6v\") pod \"612e0da4-aaaa-4050-be6d-ad14bae102be\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.910082 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ca71594-fdd4-4d9b-93b2-81f739476f9d-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.914184 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/612e0da4-aaaa-4050-be6d-ad14bae102be-logs" (OuterVolumeSpecName: "logs") pod "612e0da4-aaaa-4050-be6d-ad14bae102be" (UID: "612e0da4-aaaa-4050-be6d-ad14bae102be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.917390 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/612e0da4-aaaa-4050-be6d-ad14bae102be-kube-api-access-86r6v" (OuterVolumeSpecName: "kube-api-access-86r6v") pod "612e0da4-aaaa-4050-be6d-ad14bae102be" (UID: "612e0da4-aaaa-4050-be6d-ad14bae102be"). InnerVolumeSpecName "kube-api-access-86r6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:41 crc kubenswrapper[4894]: E1208 15:06:41.934237 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data podName:612e0da4-aaaa-4050-be6d-ad14bae102be nodeName:}" failed. No retries permitted until 2025-12-08 15:06:42.434208731 +0000 UTC m=+1223.534214846 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data") pod "612e0da4-aaaa-4050-be6d-ad14bae102be" (UID: "612e0da4-aaaa-4050-be6d-ad14bae102be") : error deleting /var/lib/kubelet/pods/612e0da4-aaaa-4050-be6d-ad14bae102be/volume-subpaths: remove /var/lib/kubelet/pods/612e0da4-aaaa-4050-be6d-ad14bae102be/volume-subpaths: no such file or directory Dec 08 15:06:41 crc kubenswrapper[4894]: I1208 15:06:41.937392 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "612e0da4-aaaa-4050-be6d-ad14bae102be" (UID: "612e0da4-aaaa-4050-be6d-ad14bae102be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.012323 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/612e0da4-aaaa-4050-be6d-ad14bae102be-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.012371 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86r6v\" (UniqueName: \"kubernetes.io/projected/612e0da4-aaaa-4050-be6d-ad14bae102be-kube-api-access-86r6v\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.012384 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.123283 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.143429 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.169400 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: E1208 15:06:42.169855 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-api" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.169875 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-api" Dec 08 15:06:42 crc kubenswrapper[4894]: E1208 15:06:42.169887 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca71594-fdd4-4d9b-93b2-81f739476f9d" containerName="nova-scheduler-scheduler" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.169895 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca71594-fdd4-4d9b-93b2-81f739476f9d" containerName="nova-scheduler-scheduler" Dec 08 15:06:42 crc kubenswrapper[4894]: E1208 15:06:42.169927 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-log" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.169933 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-log" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.170119 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-log" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.170144 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca71594-fdd4-4d9b-93b2-81f739476f9d" containerName="nova-scheduler-scheduler" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.170155 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" containerName="nova-api-api" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.170902 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.172936 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.187484 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.215831 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rbv2\" (UniqueName: \"kubernetes.io/projected/bf71a142-93b5-43a9-89f3-c33216987492-kube-api-access-7rbv2\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.215905 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-config-data\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.215939 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.318503 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rbv2\" (UniqueName: \"kubernetes.io/projected/bf71a142-93b5-43a9-89f3-c33216987492-kube-api-access-7rbv2\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.318582 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-config-data\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.318621 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.323766 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-config-data\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.341225 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.341696 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rbv2\" (UniqueName: \"kubernetes.io/projected/bf71a142-93b5-43a9-89f3-c33216987492-kube-api-access-7rbv2\") pod \"nova-scheduler-0\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.489599 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.521953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data\") pod \"612e0da4-aaaa-4050-be6d-ad14bae102be\" (UID: \"612e0da4-aaaa-4050-be6d-ad14bae102be\") " Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.525559 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data" (OuterVolumeSpecName: "config-data") pod "612e0da4-aaaa-4050-be6d-ad14bae102be" (UID: "612e0da4-aaaa-4050-be6d-ad14bae102be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.627690 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/612e0da4-aaaa-4050-be6d-ad14bae102be-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.825787 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerStarted","Data":"c297dace0c635bb19e87af52b7e3bfbb0964aea36dab11242d7a3f6e9ea31a2b"} Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.825878 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerStarted","Data":"7c259a409195ca407134a337eecec6b0e3cf0f0ef9aad4e3c9bb640435e3fa86"} Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.830016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"612e0da4-aaaa-4050-be6d-ad14bae102be","Type":"ContainerDied","Data":"e28407ed1fb3661f2d151147a973905b6a18fc09cf46e284f9c3bdb39909b354"} Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.830076 4894 scope.go:117] "RemoveContainer" containerID="252160400ed2b8328b7d9cf5aa1a27f609ca80f34770259b462f901fbd17e9fa" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.830134 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.861580 4894 scope.go:117] "RemoveContainer" containerID="1044c08a8de9fc242c3ff230045c77b9f2828ed456ba450a77b33fbc912a13a6" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.878306 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.895623 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.919634 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.922630 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.924494 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.931277 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.939312 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzcbd\" (UniqueName: \"kubernetes.io/projected/f62cb239-60aa-4741-9d00-a783ed8996e8-kube-api-access-jzcbd\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.939471 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.939517 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-config-data\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:42 crc kubenswrapper[4894]: I1208 15:06:42.939706 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62cb239-60aa-4741-9d00-a783ed8996e8-logs\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.012233 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.041730 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzcbd\" (UniqueName: \"kubernetes.io/projected/f62cb239-60aa-4741-9d00-a783ed8996e8-kube-api-access-jzcbd\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.041793 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.041845 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-config-data\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.041935 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62cb239-60aa-4741-9d00-a783ed8996e8-logs\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.042547 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62cb239-60aa-4741-9d00-a783ed8996e8-logs\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.046768 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.048103 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-config-data\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.066558 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzcbd\" (UniqueName: \"kubernetes.io/projected/f62cb239-60aa-4741-9d00-a783ed8996e8-kube-api-access-jzcbd\") pod \"nova-api-0\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.209769 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca71594-fdd4-4d9b-93b2-81f739476f9d" path="/var/lib/kubelet/pods/3ca71594-fdd4-4d9b-93b2-81f739476f9d/volumes" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.210609 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="612e0da4-aaaa-4050-be6d-ad14bae102be" path="/var/lib/kubelet/pods/612e0da4-aaaa-4050-be6d-ad14bae102be/volumes" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.244196 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.731236 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.732830 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.748085 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:06:43 crc kubenswrapper[4894]: W1208 15:06:43.751303 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62cb239_60aa_4741_9d00_a783ed8996e8.slice/crio-7aae7f83f0bcf14207214d34d01bc8fc1e552651c6a039303646063331486cdf WatchSource:0}: Error finding container 7aae7f83f0bcf14207214d34d01bc8fc1e552651c6a039303646063331486cdf: Status 404 returned error can't find the container with id 7aae7f83f0bcf14207214d34d01bc8fc1e552651c6a039303646063331486cdf Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.845424 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerStarted","Data":"dbf1d33af06b2a3e9710698e70a18c56e8c5d458c17453edbde6cebc579a0024"} Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.853517 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bf71a142-93b5-43a9-89f3-c33216987492","Type":"ContainerStarted","Data":"40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c"} Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.853766 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bf71a142-93b5-43a9-89f3-c33216987492","Type":"ContainerStarted","Data":"8ba95facb0d91d719e88efcc09907adf1bde51e72bba50d5a7f57b7e57e16fbd"} Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.857128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f62cb239-60aa-4741-9d00-a783ed8996e8","Type":"ContainerStarted","Data":"7aae7f83f0bcf14207214d34d01bc8fc1e552651c6a039303646063331486cdf"} Dec 08 15:06:43 crc kubenswrapper[4894]: I1208 15:06:43.877921 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.877894805 podStartE2EDuration="1.877894805s" podCreationTimestamp="2025-12-08 15:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:43.871425913 +0000 UTC m=+1224.971432028" watchObservedRunningTime="2025-12-08 15:06:43.877894805 +0000 UTC m=+1224.977900930" Dec 08 15:06:44 crc kubenswrapper[4894]: I1208 15:06:44.089152 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 08 15:06:44 crc kubenswrapper[4894]: I1208 15:06:44.868324 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f62cb239-60aa-4741-9d00-a783ed8996e8","Type":"ContainerStarted","Data":"ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0"} Dec 08 15:06:44 crc kubenswrapper[4894]: I1208 15:06:44.868378 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f62cb239-60aa-4741-9d00-a783ed8996e8","Type":"ContainerStarted","Data":"769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a"} Dec 08 15:06:44 crc kubenswrapper[4894]: I1208 15:06:44.872183 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerStarted","Data":"e392ac04778149d84ea11cf8c7475828632d7c2fba249e1cd62ab6b593e7647c"} Dec 08 15:06:44 crc kubenswrapper[4894]: I1208 15:06:44.893097 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.89307756 podStartE2EDuration="2.89307756s" podCreationTimestamp="2025-12-08 15:06:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:06:44.887525116 +0000 UTC m=+1225.987531231" watchObservedRunningTime="2025-12-08 15:06:44.89307756 +0000 UTC m=+1225.993083675" Dec 08 15:06:47 crc kubenswrapper[4894]: I1208 15:06:47.057447 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 08 15:06:47 crc kubenswrapper[4894]: I1208 15:06:47.490646 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 08 15:06:48 crc kubenswrapper[4894]: I1208 15:06:48.731121 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 15:06:48 crc kubenswrapper[4894]: I1208 15:06:48.731571 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 15:06:48 crc kubenswrapper[4894]: I1208 15:06:48.911708 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerStarted","Data":"1c96605e0a91c42efdbc235718bb0e4412e5ac317ec0ef72e9bdb6608c707623"} Dec 08 15:06:48 crc kubenswrapper[4894]: I1208 15:06:48.912903 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:06:48 crc kubenswrapper[4894]: I1208 15:06:48.934088 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.78032512 podStartE2EDuration="8.934071609s" podCreationTimestamp="2025-12-08 15:06:40 +0000 UTC" firstStartedPulling="2025-12-08 15:06:41.79401306 +0000 UTC m=+1222.894019185" lastFinishedPulling="2025-12-08 15:06:47.947759559 +0000 UTC m=+1229.047765674" observedRunningTime="2025-12-08 15:06:48.93280219 +0000 UTC m=+1230.032808305" watchObservedRunningTime="2025-12-08 15:06:48.934071609 +0000 UTC m=+1230.034077724" Dec 08 15:06:49 crc kubenswrapper[4894]: I1208 15:06:49.746073 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 15:06:49 crc kubenswrapper[4894]: I1208 15:06:49.746097 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 15:06:52 crc kubenswrapper[4894]: I1208 15:06:52.490137 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 08 15:06:52 crc kubenswrapper[4894]: I1208 15:06:52.520494 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 08 15:06:52 crc kubenswrapper[4894]: I1208 15:06:52.993071 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 08 15:06:53 crc kubenswrapper[4894]: I1208 15:06:53.244888 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 15:06:53 crc kubenswrapper[4894]: I1208 15:06:53.244960 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 15:06:54 crc kubenswrapper[4894]: I1208 15:06:54.328159 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 15:06:54 crc kubenswrapper[4894]: I1208 15:06:54.328295 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 08 15:06:58 crc kubenswrapper[4894]: I1208 15:06:58.737447 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 15:06:58 crc kubenswrapper[4894]: I1208 15:06:58.739093 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 15:06:58 crc kubenswrapper[4894]: I1208 15:06:58.745881 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.012912 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.736610 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.787395 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-config-data\") pod \"8228dc10-46b5-4855-a46c-aa6e16d09735\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.787429 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-combined-ca-bundle\") pod \"8228dc10-46b5-4855-a46c-aa6e16d09735\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.787636 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndk5c\" (UniqueName: \"kubernetes.io/projected/8228dc10-46b5-4855-a46c-aa6e16d09735-kube-api-access-ndk5c\") pod \"8228dc10-46b5-4855-a46c-aa6e16d09735\" (UID: \"8228dc10-46b5-4855-a46c-aa6e16d09735\") " Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.792434 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8228dc10-46b5-4855-a46c-aa6e16d09735-kube-api-access-ndk5c" (OuterVolumeSpecName: "kube-api-access-ndk5c") pod "8228dc10-46b5-4855-a46c-aa6e16d09735" (UID: "8228dc10-46b5-4855-a46c-aa6e16d09735"). InnerVolumeSpecName "kube-api-access-ndk5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.815155 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-config-data" (OuterVolumeSpecName: "config-data") pod "8228dc10-46b5-4855-a46c-aa6e16d09735" (UID: "8228dc10-46b5-4855-a46c-aa6e16d09735"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.818672 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8228dc10-46b5-4855-a46c-aa6e16d09735" (UID: "8228dc10-46b5-4855-a46c-aa6e16d09735"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.890023 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndk5c\" (UniqueName: \"kubernetes.io/projected/8228dc10-46b5-4855-a46c-aa6e16d09735-kube-api-access-ndk5c\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.890056 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:06:59 crc kubenswrapper[4894]: I1208 15:06:59.890066 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8228dc10-46b5-4855-a46c-aa6e16d09735-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.017808 4894 generic.go:334] "Generic (PLEG): container finished" podID="8228dc10-46b5-4855-a46c-aa6e16d09735" containerID="fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8" exitCode=137 Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.017866 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.017897 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8228dc10-46b5-4855-a46c-aa6e16d09735","Type":"ContainerDied","Data":"fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8"} Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.018600 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8228dc10-46b5-4855-a46c-aa6e16d09735","Type":"ContainerDied","Data":"953f1d726d0501a0c3173a373e5b5e48f3e061823b0d3e22fe859857235a06ac"} Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.018625 4894 scope.go:117] "RemoveContainer" containerID="fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.050093 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.050708 4894 scope.go:117] "RemoveContainer" containerID="fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8" Dec 08 15:07:00 crc kubenswrapper[4894]: E1208 15:07:00.051072 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8\": container with ID starting with fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8 not found: ID does not exist" containerID="fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.051100 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8"} err="failed to get container status \"fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8\": rpc error: code = NotFound desc = could not find container \"fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8\": container with ID starting with fb9a7a1135f9a12de0c7b577b4e18d74d13fff9bae3b96aac09ed0fbf636c4a8 not found: ID does not exist" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.058760 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.086901 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:07:00 crc kubenswrapper[4894]: E1208 15:07:00.088935 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8228dc10-46b5-4855-a46c-aa6e16d09735" containerName="nova-cell1-novncproxy-novncproxy" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.088964 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8228dc10-46b5-4855-a46c-aa6e16d09735" containerName="nova-cell1-novncproxy-novncproxy" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.089485 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8228dc10-46b5-4855-a46c-aa6e16d09735" containerName="nova-cell1-novncproxy-novncproxy" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.090604 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.092837 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.094735 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.096290 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.115661 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.197435 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.197544 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.197606 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.197644 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx2db\" (UniqueName: \"kubernetes.io/projected/1aee5a31-802a-4d05-b242-a7228bc76fbf-kube-api-access-cx2db\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.197664 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.299187 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.299262 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx2db\" (UniqueName: \"kubernetes.io/projected/1aee5a31-802a-4d05-b242-a7228bc76fbf-kube-api-access-cx2db\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.299291 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.299311 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.299403 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.303664 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.304151 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.304983 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.307904 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1aee5a31-802a-4d05-b242-a7228bc76fbf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.316118 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx2db\" (UniqueName: \"kubernetes.io/projected/1aee5a31-802a-4d05-b242-a7228bc76fbf-kube-api-access-cx2db\") pod \"nova-cell1-novncproxy-0\" (UID: \"1aee5a31-802a-4d05-b242-a7228bc76fbf\") " pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.425943 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:00 crc kubenswrapper[4894]: I1208 15:07:00.872558 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 08 15:07:00 crc kubenswrapper[4894]: W1208 15:07:00.874370 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1aee5a31_802a_4d05_b242_a7228bc76fbf.slice/crio-67334710e2d5cd13f61c2200d029956d39a6d6b43532852c724a150dee1728fe WatchSource:0}: Error finding container 67334710e2d5cd13f61c2200d029956d39a6d6b43532852c724a150dee1728fe: Status 404 returned error can't find the container with id 67334710e2d5cd13f61c2200d029956d39a6d6b43532852c724a150dee1728fe Dec 08 15:07:01 crc kubenswrapper[4894]: I1208 15:07:01.033896 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1aee5a31-802a-4d05-b242-a7228bc76fbf","Type":"ContainerStarted","Data":"67334710e2d5cd13f61c2200d029956d39a6d6b43532852c724a150dee1728fe"} Dec 08 15:07:01 crc kubenswrapper[4894]: I1208 15:07:01.206294 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8228dc10-46b5-4855-a46c-aa6e16d09735" path="/var/lib/kubelet/pods/8228dc10-46b5-4855-a46c-aa6e16d09735/volumes" Dec 08 15:07:02 crc kubenswrapper[4894]: I1208 15:07:02.047331 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1aee5a31-802a-4d05-b242-a7228bc76fbf","Type":"ContainerStarted","Data":"4db785abeddfc79cd7af433d1a9b814e0fe1131c007ba913b7b77228c9713daa"} Dec 08 15:07:02 crc kubenswrapper[4894]: I1208 15:07:02.072996 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.072971009 podStartE2EDuration="2.072971009s" podCreationTimestamp="2025-12-08 15:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:07:02.066338941 +0000 UTC m=+1243.166345066" watchObservedRunningTime="2025-12-08 15:07:02.072971009 +0000 UTC m=+1243.172977134" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.250662 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.250738 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.251605 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.255711 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.257922 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.258507 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.511294 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-s5t45"] Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.515221 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.533848 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-s5t45"] Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.558323 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjvpm\" (UniqueName: \"kubernetes.io/projected/de234281-8cd3-4b19-8cab-1eff756ff40c-kube-api-access-jjvpm\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.558381 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-config\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.558413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.558431 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.558444 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.558525 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.660525 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.660586 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.660699 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.660755 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjvpm\" (UniqueName: \"kubernetes.io/projected/de234281-8cd3-4b19-8cab-1eff756ff40c-kube-api-access-jjvpm\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.660796 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-config\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.660838 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.662072 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.662484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.662525 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.662586 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-config\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.662734 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.684153 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjvpm\" (UniqueName: \"kubernetes.io/projected/de234281-8cd3-4b19-8cab-1eff756ff40c-kube-api-access-jjvpm\") pod \"dnsmasq-dns-59cf4bdb65-s5t45\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:03 crc kubenswrapper[4894]: I1208 15:07:03.855778 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:04 crc kubenswrapper[4894]: I1208 15:07:04.412544 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-s5t45"] Dec 08 15:07:04 crc kubenswrapper[4894]: W1208 15:07:04.432987 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde234281_8cd3_4b19_8cab_1eff756ff40c.slice/crio-4a139b2934c5fe28396f15e0147bbe2a26ef511c1b11dad83c247b71614d8e89 WatchSource:0}: Error finding container 4a139b2934c5fe28396f15e0147bbe2a26ef511c1b11dad83c247b71614d8e89: Status 404 returned error can't find the container with id 4a139b2934c5fe28396f15e0147bbe2a26ef511c1b11dad83c247b71614d8e89 Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.091996 4894 generic.go:334] "Generic (PLEG): container finished" podID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerID="9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426" exitCode=0 Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.092095 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" event={"ID":"de234281-8cd3-4b19-8cab-1eff756ff40c","Type":"ContainerDied","Data":"9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426"} Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.092604 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" event={"ID":"de234281-8cd3-4b19-8cab-1eff756ff40c","Type":"ContainerStarted","Data":"4a139b2934c5fe28396f15e0147bbe2a26ef511c1b11dad83c247b71614d8e89"} Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.426680 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.541850 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.542368 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="proxy-httpd" containerID="cri-o://1c96605e0a91c42efdbc235718bb0e4412e5ac317ec0ef72e9bdb6608c707623" gracePeriod=30 Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.542421 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="sg-core" containerID="cri-o://e392ac04778149d84ea11cf8c7475828632d7c2fba249e1cd62ab6b593e7647c" gracePeriod=30 Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.542421 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-notification-agent" containerID="cri-o://dbf1d33af06b2a3e9710698e70a18c56e8c5d458c17453edbde6cebc579a0024" gracePeriod=30 Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.542902 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-central-agent" containerID="cri-o://c297dace0c635bb19e87af52b7e3bfbb0964aea36dab11242d7a3f6e9ea31a2b" gracePeriod=30 Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.645487 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.194:3000/\": read tcp 10.217.0.2:49770->10.217.0.194:3000: read: connection reset by peer" Dec 08 15:07:05 crc kubenswrapper[4894]: I1208 15:07:05.849496 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.104585 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" event={"ID":"de234281-8cd3-4b19-8cab-1eff756ff40c","Type":"ContainerStarted","Data":"3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653"} Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.104751 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.107523 4894 generic.go:334] "Generic (PLEG): container finished" podID="8c25b112-8906-4404-a824-e20084014c1b" containerID="1c96605e0a91c42efdbc235718bb0e4412e5ac317ec0ef72e9bdb6608c707623" exitCode=0 Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.107552 4894 generic.go:334] "Generic (PLEG): container finished" podID="8c25b112-8906-4404-a824-e20084014c1b" containerID="e392ac04778149d84ea11cf8c7475828632d7c2fba249e1cd62ab6b593e7647c" exitCode=2 Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.107564 4894 generic.go:334] "Generic (PLEG): container finished" podID="8c25b112-8906-4404-a824-e20084014c1b" containerID="c297dace0c635bb19e87af52b7e3bfbb0964aea36dab11242d7a3f6e9ea31a2b" exitCode=0 Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.107831 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-log" containerID="cri-o://769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a" gracePeriod=30 Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.108069 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerDied","Data":"1c96605e0a91c42efdbc235718bb0e4412e5ac317ec0ef72e9bdb6608c707623"} Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.108090 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerDied","Data":"e392ac04778149d84ea11cf8c7475828632d7c2fba249e1cd62ab6b593e7647c"} Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.108100 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerDied","Data":"c297dace0c635bb19e87af52b7e3bfbb0964aea36dab11242d7a3f6e9ea31a2b"} Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.108153 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-api" containerID="cri-o://ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0" gracePeriod=30 Dec 08 15:07:06 crc kubenswrapper[4894]: I1208 15:07:06.128549 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" podStartSLOduration=3.128531852 podStartE2EDuration="3.128531852s" podCreationTimestamp="2025-12-08 15:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:07:06.122268376 +0000 UTC m=+1247.222274491" watchObservedRunningTime="2025-12-08 15:07:06.128531852 +0000 UTC m=+1247.228537967" Dec 08 15:07:07 crc kubenswrapper[4894]: I1208 15:07:07.122151 4894 generic.go:334] "Generic (PLEG): container finished" podID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerID="769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a" exitCode=143 Dec 08 15:07:07 crc kubenswrapper[4894]: I1208 15:07:07.122538 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f62cb239-60aa-4741-9d00-a783ed8996e8","Type":"ContainerDied","Data":"769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a"} Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.751453 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.792856 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62cb239-60aa-4741-9d00-a783ed8996e8-logs\") pod \"f62cb239-60aa-4741-9d00-a783ed8996e8\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.793022 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzcbd\" (UniqueName: \"kubernetes.io/projected/f62cb239-60aa-4741-9d00-a783ed8996e8-kube-api-access-jzcbd\") pod \"f62cb239-60aa-4741-9d00-a783ed8996e8\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.793081 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-config-data\") pod \"f62cb239-60aa-4741-9d00-a783ed8996e8\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.793114 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-combined-ca-bundle\") pod \"f62cb239-60aa-4741-9d00-a783ed8996e8\" (UID: \"f62cb239-60aa-4741-9d00-a783ed8996e8\") " Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.802214 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f62cb239-60aa-4741-9d00-a783ed8996e8-logs" (OuterVolumeSpecName: "logs") pod "f62cb239-60aa-4741-9d00-a783ed8996e8" (UID: "f62cb239-60aa-4741-9d00-a783ed8996e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.813297 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f62cb239-60aa-4741-9d00-a783ed8996e8-kube-api-access-jzcbd" (OuterVolumeSpecName: "kube-api-access-jzcbd") pod "f62cb239-60aa-4741-9d00-a783ed8996e8" (UID: "f62cb239-60aa-4741-9d00-a783ed8996e8"). InnerVolumeSpecName "kube-api-access-jzcbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.847093 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-config-data" (OuterVolumeSpecName: "config-data") pod "f62cb239-60aa-4741-9d00-a783ed8996e8" (UID: "f62cb239-60aa-4741-9d00-a783ed8996e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.855391 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f62cb239-60aa-4741-9d00-a783ed8996e8" (UID: "f62cb239-60aa-4741-9d00-a783ed8996e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.897892 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62cb239-60aa-4741-9d00-a783ed8996e8-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.897925 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzcbd\" (UniqueName: \"kubernetes.io/projected/f62cb239-60aa-4741-9d00-a783ed8996e8-kube-api-access-jzcbd\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.897939 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:09 crc kubenswrapper[4894]: I1208 15:07:09.897947 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62cb239-60aa-4741-9d00-a783ed8996e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.179189 4894 generic.go:334] "Generic (PLEG): container finished" podID="8c25b112-8906-4404-a824-e20084014c1b" containerID="dbf1d33af06b2a3e9710698e70a18c56e8c5d458c17453edbde6cebc579a0024" exitCode=0 Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.179289 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerDied","Data":"dbf1d33af06b2a3e9710698e70a18c56e8c5d458c17453edbde6cebc579a0024"} Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.187703 4894 generic.go:334] "Generic (PLEG): container finished" podID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerID="ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0" exitCode=0 Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.187749 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f62cb239-60aa-4741-9d00-a783ed8996e8","Type":"ContainerDied","Data":"ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0"} Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.187776 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f62cb239-60aa-4741-9d00-a783ed8996e8","Type":"ContainerDied","Data":"7aae7f83f0bcf14207214d34d01bc8fc1e552651c6a039303646063331486cdf"} Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.187792 4894 scope.go:117] "RemoveContainer" containerID="ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.187791 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.239522 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.250403 4894 scope.go:117] "RemoveContainer" containerID="769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.285784 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.315007 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:10 crc kubenswrapper[4894]: E1208 15:07:10.316204 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-api" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.316221 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-api" Dec 08 15:07:10 crc kubenswrapper[4894]: E1208 15:07:10.316287 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-log" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.316299 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-log" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.316701 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-api" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.316759 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" containerName="nova-api-log" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.318680 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.324753 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.325248 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.325364 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.366623 4894 scope.go:117] "RemoveContainer" containerID="ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0" Dec 08 15:07:10 crc kubenswrapper[4894]: E1208 15:07:10.367218 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0\": container with ID starting with ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0 not found: ID does not exist" containerID="ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.367255 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0"} err="failed to get container status \"ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0\": rpc error: code = NotFound desc = could not find container \"ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0\": container with ID starting with ba10f77d65b7bad630e5d8974de9493bf438c82e1c6ccc8d6fcafd0c29d9c3a0 not found: ID does not exist" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.367276 4894 scope.go:117] "RemoveContainer" containerID="769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a" Dec 08 15:07:10 crc kubenswrapper[4894]: E1208 15:07:10.369252 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a\": container with ID starting with 769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a not found: ID does not exist" containerID="769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.369318 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a"} err="failed to get container status \"769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a\": rpc error: code = NotFound desc = could not find container \"769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a\": container with ID starting with 769207a0bf7c00deb82967f5cb36a567e1ace4bac82df844bb6e96994906746a not found: ID does not exist" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.382663 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.409051 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.409150 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-config-data\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.409322 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-public-tls-certs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.409382 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14461434-583e-40a3-93fe-7bf70a6a05a7-logs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.409431 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99jgw\" (UniqueName: \"kubernetes.io/projected/14461434-583e-40a3-93fe-7bf70a6a05a7-kube-api-access-99jgw\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.409506 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.427054 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.450736 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.498304 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.511466 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-ceilometer-tls-certs\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.511790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-log-httpd\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.512072 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-config-data\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.512449 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.512792 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-run-httpd\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.512965 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-scripts\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.513025 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.513114 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-combined-ca-bundle\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.513213 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-sg-core-conf-yaml\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.513422 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8mg6\" (UniqueName: \"kubernetes.io/projected/8c25b112-8906-4404-a824-e20084014c1b-kube-api-access-r8mg6\") pod \"8c25b112-8906-4404-a824-e20084014c1b\" (UID: \"8c25b112-8906-4404-a824-e20084014c1b\") " Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.514952 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.515043 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-config-data\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.515609 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-public-tls-certs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.515758 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14461434-583e-40a3-93fe-7bf70a6a05a7-logs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.515840 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99jgw\" (UniqueName: \"kubernetes.io/projected/14461434-583e-40a3-93fe-7bf70a6a05a7-kube-api-access-99jgw\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.516062 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.516396 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14461434-583e-40a3-93fe-7bf70a6a05a7-logs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.519237 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.519277 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c25b112-8906-4404-a824-e20084014c1b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.523747 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-scripts" (OuterVolumeSpecName: "scripts") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.524845 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-public-tls-certs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.533122 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c25b112-8906-4404-a824-e20084014c1b-kube-api-access-r8mg6" (OuterVolumeSpecName: "kube-api-access-r8mg6") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "kube-api-access-r8mg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.537113 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.538636 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.539240 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-config-data\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.554596 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99jgw\" (UniqueName: \"kubernetes.io/projected/14461434-583e-40a3-93fe-7bf70a6a05a7-kube-api-access-99jgw\") pod \"nova-api-0\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " pod="openstack/nova-api-0" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.565882 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.592783 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.620680 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.620736 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8mg6\" (UniqueName: \"kubernetes.io/projected/8c25b112-8906-4404-a824-e20084014c1b-kube-api-access-r8mg6\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.620756 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.620767 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.637959 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.657357 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-config-data" (OuterVolumeSpecName: "config-data") pod "8c25b112-8906-4404-a824-e20084014c1b" (UID: "8c25b112-8906-4404-a824-e20084014c1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.723428 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.723463 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c25b112-8906-4404-a824-e20084014c1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:10 crc kubenswrapper[4894]: I1208 15:07:10.797209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.201125 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.206537 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f62cb239-60aa-4741-9d00-a783ed8996e8" path="/var/lib/kubelet/pods/f62cb239-60aa-4741-9d00-a783ed8996e8/volumes" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.207317 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c25b112-8906-4404-a824-e20084014c1b","Type":"ContainerDied","Data":"7c259a409195ca407134a337eecec6b0e3cf0f0ef9aad4e3c9bb640435e3fa86"} Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.207365 4894 scope.go:117] "RemoveContainer" containerID="1c96605e0a91c42efdbc235718bb0e4412e5ac317ec0ef72e9bdb6608c707623" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.221114 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.239120 4894 scope.go:117] "RemoveContainer" containerID="e392ac04778149d84ea11cf8c7475828632d7c2fba249e1cd62ab6b593e7647c" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.267268 4894 scope.go:117] "RemoveContainer" containerID="dbf1d33af06b2a3e9710698e70a18c56e8c5d458c17453edbde6cebc579a0024" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.286511 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.309121 4894 scope.go:117] "RemoveContainer" containerID="c297dace0c635bb19e87af52b7e3bfbb0964aea36dab11242d7a3f6e9ea31a2b" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.312392 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.353920 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.367328 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:07:11 crc kubenswrapper[4894]: E1208 15:07:11.367771 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-central-agent" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.367783 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-central-agent" Dec 08 15:07:11 crc kubenswrapper[4894]: E1208 15:07:11.367802 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-notification-agent" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.367808 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-notification-agent" Dec 08 15:07:11 crc kubenswrapper[4894]: E1208 15:07:11.367834 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="proxy-httpd" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.367842 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="proxy-httpd" Dec 08 15:07:11 crc kubenswrapper[4894]: E1208 15:07:11.367865 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="sg-core" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.367871 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="sg-core" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.368048 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-notification-agent" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.368064 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="sg-core" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.368071 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="proxy-httpd" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.368086 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c25b112-8906-4404-a824-e20084014c1b" containerName="ceilometer-central-agent" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.376316 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.379034 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.383798 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.384200 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.384438 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.441475 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.441535 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbjj\" (UniqueName: \"kubernetes.io/projected/4463b2d0-71f4-4554-a73d-7ea950f87081-kube-api-access-vxbjj\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.442039 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4463b2d0-71f4-4554-a73d-7ea950f87081-log-httpd\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.442130 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-scripts\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.442166 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.442296 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4463b2d0-71f4-4554-a73d-7ea950f87081-run-httpd\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.442349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-config-data\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.442391 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.517031 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-lkqq6"] Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.519145 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.523144 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.523333 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.536599 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lkqq6"] Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.544797 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-scripts\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.544914 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4463b2d0-71f4-4554-a73d-7ea950f87081-log-httpd\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.544977 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-scripts\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545009 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545065 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4463b2d0-71f4-4554-a73d-7ea950f87081-run-httpd\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545107 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-config-data\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545146 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-config-data\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545191 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545277 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scwdh\" (UniqueName: \"kubernetes.io/projected/f846ca80-7b63-4a22-b639-768158cd85c3-kube-api-access-scwdh\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545298 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4463b2d0-71f4-4554-a73d-7ea950f87081-log-httpd\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545320 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545352 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbjj\" (UniqueName: \"kubernetes.io/projected/4463b2d0-71f4-4554-a73d-7ea950f87081-kube-api-access-vxbjj\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.545395 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.546015 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4463b2d0-71f4-4554-a73d-7ea950f87081-run-httpd\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.556120 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.556236 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.556397 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-config-data\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.556715 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-scripts\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.560462 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4463b2d0-71f4-4554-a73d-7ea950f87081-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.567192 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbjj\" (UniqueName: \"kubernetes.io/projected/4463b2d0-71f4-4554-a73d-7ea950f87081-kube-api-access-vxbjj\") pod \"ceilometer-0\" (UID: \"4463b2d0-71f4-4554-a73d-7ea950f87081\") " pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.646669 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-scripts\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.646794 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-config-data\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.646965 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scwdh\" (UniqueName: \"kubernetes.io/projected/f846ca80-7b63-4a22-b639-768158cd85c3-kube-api-access-scwdh\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.647038 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.654290 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-scripts\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.654543 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.655190 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-config-data\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.665755 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scwdh\" (UniqueName: \"kubernetes.io/projected/f846ca80-7b63-4a22-b639-768158cd85c3-kube-api-access-scwdh\") pod \"nova-cell1-cell-mapping-lkqq6\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.838239 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 08 15:07:11 crc kubenswrapper[4894]: I1208 15:07:11.852356 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:12 crc kubenswrapper[4894]: I1208 15:07:12.212283 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14461434-583e-40a3-93fe-7bf70a6a05a7","Type":"ContainerStarted","Data":"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf"} Dec 08 15:07:12 crc kubenswrapper[4894]: I1208 15:07:12.212607 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14461434-583e-40a3-93fe-7bf70a6a05a7","Type":"ContainerStarted","Data":"2ea7d7827c4291282501eca5845a81651e4d2d339369a3e4304fbf5ba436e16d"} Dec 08 15:07:12 crc kubenswrapper[4894]: I1208 15:07:12.373804 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 08 15:07:12 crc kubenswrapper[4894]: W1208 15:07:12.377799 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4463b2d0_71f4_4554_a73d_7ea950f87081.slice/crio-7ba96a602289f927854c73d77df18269d702e14f96d668975890a58879498d8c WatchSource:0}: Error finding container 7ba96a602289f927854c73d77df18269d702e14f96d668975890a58879498d8c: Status 404 returned error can't find the container with id 7ba96a602289f927854c73d77df18269d702e14f96d668975890a58879498d8c Dec 08 15:07:12 crc kubenswrapper[4894]: I1208 15:07:12.380185 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:07:12 crc kubenswrapper[4894]: I1208 15:07:12.458354 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lkqq6"] Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.209261 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c25b112-8906-4404-a824-e20084014c1b" path="/var/lib/kubelet/pods/8c25b112-8906-4404-a824-e20084014c1b/volumes" Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.228229 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14461434-583e-40a3-93fe-7bf70a6a05a7","Type":"ContainerStarted","Data":"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2"} Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.230203 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lkqq6" event={"ID":"f846ca80-7b63-4a22-b639-768158cd85c3","Type":"ContainerStarted","Data":"9b21b545fd572ccb99efe6565300d9a53548901f69ecdf10bf88f6535ed32a37"} Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.230270 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lkqq6" event={"ID":"f846ca80-7b63-4a22-b639-768158cd85c3","Type":"ContainerStarted","Data":"5c9539df4524d2ec6b1cef5fca97303d292887c7d725f5d9fb4ab98a3c0b8af3"} Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.233102 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4463b2d0-71f4-4554-a73d-7ea950f87081","Type":"ContainerStarted","Data":"41bd891612322503b80ffca5c043b1cd29e65b305a37a212a0c1db4fbcf6e85b"} Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.233143 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4463b2d0-71f4-4554-a73d-7ea950f87081","Type":"ContainerStarted","Data":"7ba96a602289f927854c73d77df18269d702e14f96d668975890a58879498d8c"} Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.252269 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.252255152 podStartE2EDuration="3.252255152s" podCreationTimestamp="2025-12-08 15:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:07:13.251117376 +0000 UTC m=+1254.351123501" watchObservedRunningTime="2025-12-08 15:07:13.252255152 +0000 UTC m=+1254.352261267" Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.282536 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-lkqq6" podStartSLOduration=2.282499387 podStartE2EDuration="2.282499387s" podCreationTimestamp="2025-12-08 15:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:07:13.279250145 +0000 UTC m=+1254.379256280" watchObservedRunningTime="2025-12-08 15:07:13.282499387 +0000 UTC m=+1254.382505502" Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.859053 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.951416 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-w87xf"] Dec 08 15:07:13 crc kubenswrapper[4894]: I1208 15:07:13.951659 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerName="dnsmasq-dns" containerID="cri-o://9a06884443733ed8538c8804af17f12438fa67ec5fae8fc3df0592d909115676" gracePeriod=10 Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.089382 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.187:5353: connect: connection refused" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.249395 4894 generic.go:334] "Generic (PLEG): container finished" podID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerID="9a06884443733ed8538c8804af17f12438fa67ec5fae8fc3df0592d909115676" exitCode=0 Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.249602 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" event={"ID":"0c18cbf5-83bf-424d-9870-1f921b25a9e2","Type":"ContainerDied","Data":"9a06884443733ed8538c8804af17f12438fa67ec5fae8fc3df0592d909115676"} Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.256020 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4463b2d0-71f4-4554-a73d-7ea950f87081","Type":"ContainerStarted","Data":"6a027f1f9f56bacc5e81df13b2184c000d5f6739f751a96d2f525f63626dc365"} Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.552505 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.743381 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-svc\") pod \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.743455 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdn2n\" (UniqueName: \"kubernetes.io/projected/0c18cbf5-83bf-424d-9870-1f921b25a9e2-kube-api-access-sdn2n\") pod \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.743577 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-nb\") pod \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.743648 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-swift-storage-0\") pod \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.743728 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-config\") pod \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.743855 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-sb\") pod \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\" (UID: \"0c18cbf5-83bf-424d-9870-1f921b25a9e2\") " Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.754994 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c18cbf5-83bf-424d-9870-1f921b25a9e2-kube-api-access-sdn2n" (OuterVolumeSpecName: "kube-api-access-sdn2n") pod "0c18cbf5-83bf-424d-9870-1f921b25a9e2" (UID: "0c18cbf5-83bf-424d-9870-1f921b25a9e2"). InnerVolumeSpecName "kube-api-access-sdn2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.806022 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c18cbf5-83bf-424d-9870-1f921b25a9e2" (UID: "0c18cbf5-83bf-424d-9870-1f921b25a9e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.815747 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c18cbf5-83bf-424d-9870-1f921b25a9e2" (UID: "0c18cbf5-83bf-424d-9870-1f921b25a9e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.817178 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c18cbf5-83bf-424d-9870-1f921b25a9e2" (UID: "0c18cbf5-83bf-424d-9870-1f921b25a9e2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.817876 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0c18cbf5-83bf-424d-9870-1f921b25a9e2" (UID: "0c18cbf5-83bf-424d-9870-1f921b25a9e2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.831734 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-config" (OuterVolumeSpecName: "config") pod "0c18cbf5-83bf-424d-9870-1f921b25a9e2" (UID: "0c18cbf5-83bf-424d-9870-1f921b25a9e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.849961 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.850137 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.850217 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.850305 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdn2n\" (UniqueName: \"kubernetes.io/projected/0c18cbf5-83bf-424d-9870-1f921b25a9e2-kube-api-access-sdn2n\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.850361 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:14 crc kubenswrapper[4894]: I1208 15:07:14.850423 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c18cbf5-83bf-424d-9870-1f921b25a9e2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:15 crc kubenswrapper[4894]: I1208 15:07:15.267098 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4463b2d0-71f4-4554-a73d-7ea950f87081","Type":"ContainerStarted","Data":"661c9e4730db1fd124801c7fc319a48f3b0752aa888c4ef18e2c672db425bd29"} Dec 08 15:07:15 crc kubenswrapper[4894]: I1208 15:07:15.268860 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" event={"ID":"0c18cbf5-83bf-424d-9870-1f921b25a9e2","Type":"ContainerDied","Data":"078f6a58921091195e9969e36d66805ae99e3a606b577357d610ffcc74780a72"} Dec 08 15:07:15 crc kubenswrapper[4894]: I1208 15:07:15.268907 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-w87xf" Dec 08 15:07:15 crc kubenswrapper[4894]: I1208 15:07:15.268918 4894 scope.go:117] "RemoveContainer" containerID="9a06884443733ed8538c8804af17f12438fa67ec5fae8fc3df0592d909115676" Dec 08 15:07:15 crc kubenswrapper[4894]: I1208 15:07:15.296566 4894 scope.go:117] "RemoveContainer" containerID="2bbf1942037b9d75778f63717d403ec066afc3594475ee2d3d4e143903706269" Dec 08 15:07:15 crc kubenswrapper[4894]: I1208 15:07:15.306396 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-w87xf"] Dec 08 15:07:15 crc kubenswrapper[4894]: I1208 15:07:15.312964 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-w87xf"] Dec 08 15:07:16 crc kubenswrapper[4894]: I1208 15:07:16.281601 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4463b2d0-71f4-4554-a73d-7ea950f87081","Type":"ContainerStarted","Data":"935ab18b7a9f6b116c2a0803011938dd773bb4c87a15f4a2f804800195a4aaca"} Dec 08 15:07:16 crc kubenswrapper[4894]: I1208 15:07:16.282058 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 08 15:07:17 crc kubenswrapper[4894]: I1208 15:07:17.207242 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" path="/var/lib/kubelet/pods/0c18cbf5-83bf-424d-9870-1f921b25a9e2/volumes" Dec 08 15:07:18 crc kubenswrapper[4894]: I1208 15:07:18.302708 4894 generic.go:334] "Generic (PLEG): container finished" podID="f846ca80-7b63-4a22-b639-768158cd85c3" containerID="9b21b545fd572ccb99efe6565300d9a53548901f69ecdf10bf88f6535ed32a37" exitCode=0 Dec 08 15:07:18 crc kubenswrapper[4894]: I1208 15:07:18.303110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lkqq6" event={"ID":"f846ca80-7b63-4a22-b639-768158cd85c3","Type":"ContainerDied","Data":"9b21b545fd572ccb99efe6565300d9a53548901f69ecdf10bf88f6535ed32a37"} Dec 08 15:07:18 crc kubenswrapper[4894]: I1208 15:07:18.324535 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.360849458 podStartE2EDuration="7.324517049s" podCreationTimestamp="2025-12-08 15:07:11 +0000 UTC" firstStartedPulling="2025-12-08 15:07:12.379788574 +0000 UTC m=+1253.479794689" lastFinishedPulling="2025-12-08 15:07:15.343456165 +0000 UTC m=+1256.443462280" observedRunningTime="2025-12-08 15:07:16.304211151 +0000 UTC m=+1257.404217266" watchObservedRunningTime="2025-12-08 15:07:18.324517049 +0000 UTC m=+1259.424523164" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.694038 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.844789 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-scripts\") pod \"f846ca80-7b63-4a22-b639-768158cd85c3\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.845313 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-config-data\") pod \"f846ca80-7b63-4a22-b639-768158cd85c3\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.845348 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-combined-ca-bundle\") pod \"f846ca80-7b63-4a22-b639-768158cd85c3\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.845369 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scwdh\" (UniqueName: \"kubernetes.io/projected/f846ca80-7b63-4a22-b639-768158cd85c3-kube-api-access-scwdh\") pod \"f846ca80-7b63-4a22-b639-768158cd85c3\" (UID: \"f846ca80-7b63-4a22-b639-768158cd85c3\") " Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.852103 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-scripts" (OuterVolumeSpecName: "scripts") pod "f846ca80-7b63-4a22-b639-768158cd85c3" (UID: "f846ca80-7b63-4a22-b639-768158cd85c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.857514 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f846ca80-7b63-4a22-b639-768158cd85c3-kube-api-access-scwdh" (OuterVolumeSpecName: "kube-api-access-scwdh") pod "f846ca80-7b63-4a22-b639-768158cd85c3" (UID: "f846ca80-7b63-4a22-b639-768158cd85c3"). InnerVolumeSpecName "kube-api-access-scwdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.877761 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-config-data" (OuterVolumeSpecName: "config-data") pod "f846ca80-7b63-4a22-b639-768158cd85c3" (UID: "f846ca80-7b63-4a22-b639-768158cd85c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.880505 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f846ca80-7b63-4a22-b639-768158cd85c3" (UID: "f846ca80-7b63-4a22-b639-768158cd85c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.947229 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.947263 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scwdh\" (UniqueName: \"kubernetes.io/projected/f846ca80-7b63-4a22-b639-768158cd85c3-kube-api-access-scwdh\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.947276 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-scripts\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:19 crc kubenswrapper[4894]: I1208 15:07:19.947285 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f846ca80-7b63-4a22-b639-768158cd85c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.351636 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lkqq6" event={"ID":"f846ca80-7b63-4a22-b639-768158cd85c3","Type":"ContainerDied","Data":"5c9539df4524d2ec6b1cef5fca97303d292887c7d725f5d9fb4ab98a3c0b8af3"} Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.351676 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c9539df4524d2ec6b1cef5fca97303d292887c7d725f5d9fb4ab98a3c0b8af3" Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.351923 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lkqq6" Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.565473 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.565759 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="bf71a142-93b5-43a9-89f3-c33216987492" containerName="nova-scheduler-scheduler" containerID="cri-o://40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c" gracePeriod=30 Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.595940 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.596328 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-log" containerID="cri-o://54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf" gracePeriod=30 Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.596500 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-api" containerID="cri-o://a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2" gracePeriod=30 Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.611936 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.613026 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-log" containerID="cri-o://d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa" gracePeriod=30 Dec 08 15:07:20 crc kubenswrapper[4894]: I1208 15:07:20.613117 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-metadata" containerID="cri-o://0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b" gracePeriod=30 Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.301849 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.366699 4894 generic.go:334] "Generic (PLEG): container finished" podID="c42e42b6-4389-4938-973d-62cfc12cb948" containerID="d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa" exitCode=143 Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.366791 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c42e42b6-4389-4938-973d-62cfc12cb948","Type":"ContainerDied","Data":"d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa"} Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.369039 4894 generic.go:334] "Generic (PLEG): container finished" podID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerID="a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2" exitCode=0 Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.369078 4894 generic.go:334] "Generic (PLEG): container finished" podID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerID="54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf" exitCode=143 Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.369100 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14461434-583e-40a3-93fe-7bf70a6a05a7","Type":"ContainerDied","Data":"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2"} Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.369125 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14461434-583e-40a3-93fe-7bf70a6a05a7","Type":"ContainerDied","Data":"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf"} Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.369135 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"14461434-583e-40a3-93fe-7bf70a6a05a7","Type":"ContainerDied","Data":"2ea7d7827c4291282501eca5845a81651e4d2d339369a3e4304fbf5ba436e16d"} Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.369156 4894 scope.go:117] "RemoveContainer" containerID="a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.369332 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.399352 4894 scope.go:117] "RemoveContainer" containerID="54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.428665 4894 scope.go:117] "RemoveContainer" containerID="a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2" Dec 08 15:07:21 crc kubenswrapper[4894]: E1208 15:07:21.429374 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2\": container with ID starting with a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2 not found: ID does not exist" containerID="a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.429467 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2"} err="failed to get container status \"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2\": rpc error: code = NotFound desc = could not find container \"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2\": container with ID starting with a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2 not found: ID does not exist" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.429518 4894 scope.go:117] "RemoveContainer" containerID="54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf" Dec 08 15:07:21 crc kubenswrapper[4894]: E1208 15:07:21.430293 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf\": container with ID starting with 54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf not found: ID does not exist" containerID="54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.430368 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf"} err="failed to get container status \"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf\": rpc error: code = NotFound desc = could not find container \"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf\": container with ID starting with 54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf not found: ID does not exist" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.430419 4894 scope.go:117] "RemoveContainer" containerID="a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.431077 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2"} err="failed to get container status \"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2\": rpc error: code = NotFound desc = could not find container \"a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2\": container with ID starting with a582609e803e41b32dda575fe01bc73f999bbeb86c232d90f6697808842c98c2 not found: ID does not exist" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.431101 4894 scope.go:117] "RemoveContainer" containerID="54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.431480 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf"} err="failed to get container status \"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf\": rpc error: code = NotFound desc = could not find container \"54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf\": container with ID starting with 54414899059c86c8d67fc6f4e72eb1566d2c21c87fb9caff618b6e8c96b505bf not found: ID does not exist" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.475869 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99jgw\" (UniqueName: \"kubernetes.io/projected/14461434-583e-40a3-93fe-7bf70a6a05a7-kube-api-access-99jgw\") pod \"14461434-583e-40a3-93fe-7bf70a6a05a7\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.476085 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-internal-tls-certs\") pod \"14461434-583e-40a3-93fe-7bf70a6a05a7\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.476133 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-combined-ca-bundle\") pod \"14461434-583e-40a3-93fe-7bf70a6a05a7\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.476187 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-config-data\") pod \"14461434-583e-40a3-93fe-7bf70a6a05a7\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.476212 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-public-tls-certs\") pod \"14461434-583e-40a3-93fe-7bf70a6a05a7\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.476431 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14461434-583e-40a3-93fe-7bf70a6a05a7-logs\") pod \"14461434-583e-40a3-93fe-7bf70a6a05a7\" (UID: \"14461434-583e-40a3-93fe-7bf70a6a05a7\") " Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.477573 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14461434-583e-40a3-93fe-7bf70a6a05a7-logs" (OuterVolumeSpecName: "logs") pod "14461434-583e-40a3-93fe-7bf70a6a05a7" (UID: "14461434-583e-40a3-93fe-7bf70a6a05a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.485422 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14461434-583e-40a3-93fe-7bf70a6a05a7-kube-api-access-99jgw" (OuterVolumeSpecName: "kube-api-access-99jgw") pod "14461434-583e-40a3-93fe-7bf70a6a05a7" (UID: "14461434-583e-40a3-93fe-7bf70a6a05a7"). InnerVolumeSpecName "kube-api-access-99jgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.519059 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14461434-583e-40a3-93fe-7bf70a6a05a7" (UID: "14461434-583e-40a3-93fe-7bf70a6a05a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.521741 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-config-data" (OuterVolumeSpecName: "config-data") pod "14461434-583e-40a3-93fe-7bf70a6a05a7" (UID: "14461434-583e-40a3-93fe-7bf70a6a05a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.545553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "14461434-583e-40a3-93fe-7bf70a6a05a7" (UID: "14461434-583e-40a3-93fe-7bf70a6a05a7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.549000 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "14461434-583e-40a3-93fe-7bf70a6a05a7" (UID: "14461434-583e-40a3-93fe-7bf70a6a05a7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.578722 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14461434-583e-40a3-93fe-7bf70a6a05a7-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.578766 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99jgw\" (UniqueName: \"kubernetes.io/projected/14461434-583e-40a3-93fe-7bf70a6a05a7-kube-api-access-99jgw\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.578779 4894 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.578788 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.578796 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.578805 4894 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/14461434-583e-40a3-93fe-7bf70a6a05a7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.732496 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.749118 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.810412 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:21 crc kubenswrapper[4894]: E1208 15:07:21.811934 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f846ca80-7b63-4a22-b639-768158cd85c3" containerName="nova-manage" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.811981 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f846ca80-7b63-4a22-b639-768158cd85c3" containerName="nova-manage" Dec 08 15:07:21 crc kubenswrapper[4894]: E1208 15:07:21.812015 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-api" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812027 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-api" Dec 08 15:07:21 crc kubenswrapper[4894]: E1208 15:07:21.812041 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerName="dnsmasq-dns" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812058 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerName="dnsmasq-dns" Dec 08 15:07:21 crc kubenswrapper[4894]: E1208 15:07:21.812139 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-log" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812152 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-log" Dec 08 15:07:21 crc kubenswrapper[4894]: E1208 15:07:21.812182 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerName="init" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812191 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerName="init" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812756 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-log" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812829 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f846ca80-7b63-4a22-b639-768158cd85c3" containerName="nova-manage" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812877 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" containerName="nova-api-api" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.812896 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c18cbf5-83bf-424d-9870-1f921b25a9e2" containerName="dnsmasq-dns" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.819246 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.826923 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.829960 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.830539 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 08 15:07:21 crc kubenswrapper[4894]: I1208 15:07:21.831239 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.009549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.009963 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.010040 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-config-data\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.010140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-public-tls-certs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.010179 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjq6j\" (UniqueName: \"kubernetes.io/projected/635376cd-70d9-4108-942a-927f52ce27cb-kube-api-access-xjq6j\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.010220 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635376cd-70d9-4108-942a-927f52ce27cb-logs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.113450 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjq6j\" (UniqueName: \"kubernetes.io/projected/635376cd-70d9-4108-942a-927f52ce27cb-kube-api-access-xjq6j\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.113556 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635376cd-70d9-4108-942a-927f52ce27cb-logs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.113621 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.113640 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.113741 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-config-data\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.113833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-public-tls-certs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.117874 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/635376cd-70d9-4108-942a-927f52ce27cb-logs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.118575 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-public-tls-certs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.120265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.122769 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.125226 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/635376cd-70d9-4108-942a-927f52ce27cb-config-data\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.140959 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjq6j\" (UniqueName: \"kubernetes.io/projected/635376cd-70d9-4108-942a-927f52ce27cb-kube-api-access-xjq6j\") pod \"nova-api-0\" (UID: \"635376cd-70d9-4108-942a-927f52ce27cb\") " pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.145221 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.279201 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.387049 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.387111 4894 generic.go:334] "Generic (PLEG): container finished" podID="bf71a142-93b5-43a9-89f3-c33216987492" containerID="40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c" exitCode=0 Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.387218 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bf71a142-93b5-43a9-89f3-c33216987492","Type":"ContainerDied","Data":"40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c"} Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.387254 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bf71a142-93b5-43a9-89f3-c33216987492","Type":"ContainerDied","Data":"8ba95facb0d91d719e88efcc09907adf1bde51e72bba50d5a7f57b7e57e16fbd"} Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.387274 4894 scope.go:117] "RemoveContainer" containerID="40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.414153 4894 scope.go:117] "RemoveContainer" containerID="40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c" Dec 08 15:07:22 crc kubenswrapper[4894]: E1208 15:07:22.414671 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c\": container with ID starting with 40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c not found: ID does not exist" containerID="40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.414703 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c"} err="failed to get container status \"40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c\": rpc error: code = NotFound desc = could not find container \"40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c\": container with ID starting with 40714a85f5eb2a653d2ce4521c93282470d0c72add594c7c9c7aa231c76acb7c not found: ID does not exist" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.419375 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rbv2\" (UniqueName: \"kubernetes.io/projected/bf71a142-93b5-43a9-89f3-c33216987492-kube-api-access-7rbv2\") pod \"bf71a142-93b5-43a9-89f3-c33216987492\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.420552 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-config-data\") pod \"bf71a142-93b5-43a9-89f3-c33216987492\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.420609 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-combined-ca-bundle\") pod \"bf71a142-93b5-43a9-89f3-c33216987492\" (UID: \"bf71a142-93b5-43a9-89f3-c33216987492\") " Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.433968 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf71a142-93b5-43a9-89f3-c33216987492-kube-api-access-7rbv2" (OuterVolumeSpecName: "kube-api-access-7rbv2") pod "bf71a142-93b5-43a9-89f3-c33216987492" (UID: "bf71a142-93b5-43a9-89f3-c33216987492"). InnerVolumeSpecName "kube-api-access-7rbv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.461382 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf71a142-93b5-43a9-89f3-c33216987492" (UID: "bf71a142-93b5-43a9-89f3-c33216987492"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.466008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-config-data" (OuterVolumeSpecName: "config-data") pod "bf71a142-93b5-43a9-89f3-c33216987492" (UID: "bf71a142-93b5-43a9-89f3-c33216987492"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.522787 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.522844 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf71a142-93b5-43a9-89f3-c33216987492-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.522861 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rbv2\" (UniqueName: \"kubernetes.io/projected/bf71a142-93b5-43a9-89f3-c33216987492-kube-api-access-7rbv2\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.652393 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 08 15:07:22 crc kubenswrapper[4894]: W1208 15:07:22.655474 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod635376cd_70d9_4108_942a_927f52ce27cb.slice/crio-f416284b051ab17800d351e762e21fa554babefb6f67110f3e6788a0428752f9 WatchSource:0}: Error finding container f416284b051ab17800d351e762e21fa554babefb6f67110f3e6788a0428752f9: Status 404 returned error can't find the container with id f416284b051ab17800d351e762e21fa554babefb6f67110f3e6788a0428752f9 Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.723414 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.735842 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.749467 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:07:22 crc kubenswrapper[4894]: E1208 15:07:22.749880 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf71a142-93b5-43a9-89f3-c33216987492" containerName="nova-scheduler-scheduler" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.749896 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf71a142-93b5-43a9-89f3-c33216987492" containerName="nova-scheduler-scheduler" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.750079 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf71a142-93b5-43a9-89f3-c33216987492" containerName="nova-scheduler-scheduler" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.750660 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.752828 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.766718 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.929774 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc76d242-0072-4bef-a299-0a88624cb734-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.930302 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc76d242-0072-4bef-a299-0a88624cb734-config-data\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:22 crc kubenswrapper[4894]: I1208 15:07:22.930369 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccn2n\" (UniqueName: \"kubernetes.io/projected/fc76d242-0072-4bef-a299-0a88624cb734-kube-api-access-ccn2n\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.032799 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc76d242-0072-4bef-a299-0a88624cb734-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.032912 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc76d242-0072-4bef-a299-0a88624cb734-config-data\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.032976 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccn2n\" (UniqueName: \"kubernetes.io/projected/fc76d242-0072-4bef-a299-0a88624cb734-kube-api-access-ccn2n\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.038865 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc76d242-0072-4bef-a299-0a88624cb734-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.045736 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc76d242-0072-4bef-a299-0a88624cb734-config-data\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.049728 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccn2n\" (UniqueName: \"kubernetes.io/projected/fc76d242-0072-4bef-a299-0a88624cb734-kube-api-access-ccn2n\") pod \"nova-scheduler-0\" (UID: \"fc76d242-0072-4bef-a299-0a88624cb734\") " pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.184483 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.208483 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14461434-583e-40a3-93fe-7bf70a6a05a7" path="/var/lib/kubelet/pods/14461434-583e-40a3-93fe-7bf70a6a05a7/volumes" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.209132 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf71a142-93b5-43a9-89f3-c33216987492" path="/var/lib/kubelet/pods/bf71a142-93b5-43a9-89f3-c33216987492/volumes" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.412131 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"635376cd-70d9-4108-942a-927f52ce27cb","Type":"ContainerStarted","Data":"d7071dd7d2e6d43bf6bf85800f3fa012fded78e2968919ebc04978947a45a485"} Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.412496 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"635376cd-70d9-4108-942a-927f52ce27cb","Type":"ContainerStarted","Data":"e96e45ca8cdfdf902487e9309d720bff71dee0e7490020af608b6f8caf533aca"} Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.412513 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"635376cd-70d9-4108-942a-927f52ce27cb","Type":"ContainerStarted","Data":"f416284b051ab17800d351e762e21fa554babefb6f67110f3e6788a0428752f9"} Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.449149 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.449126352 podStartE2EDuration="2.449126352s" podCreationTimestamp="2025-12-08 15:07:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:07:23.433039469 +0000 UTC m=+1264.533045604" watchObservedRunningTime="2025-12-08 15:07:23.449126352 +0000 UTC m=+1264.549132467" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.653740 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.747779 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:45502->10.217.0.192:8775: read: connection reset by peer" Dec 08 15:07:23 crc kubenswrapper[4894]: I1208 15:07:23.747849 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:45488->10.217.0.192:8775: read: connection reset by peer" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.275636 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.366796 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wztfv\" (UniqueName: \"kubernetes.io/projected/c42e42b6-4389-4938-973d-62cfc12cb948-kube-api-access-wztfv\") pod \"c42e42b6-4389-4938-973d-62cfc12cb948\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.366935 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-config-data\") pod \"c42e42b6-4389-4938-973d-62cfc12cb948\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.367963 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-nova-metadata-tls-certs\") pod \"c42e42b6-4389-4938-973d-62cfc12cb948\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.368009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-combined-ca-bundle\") pod \"c42e42b6-4389-4938-973d-62cfc12cb948\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.368081 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42e42b6-4389-4938-973d-62cfc12cb948-logs\") pod \"c42e42b6-4389-4938-973d-62cfc12cb948\" (UID: \"c42e42b6-4389-4938-973d-62cfc12cb948\") " Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.369888 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c42e42b6-4389-4938-973d-62cfc12cb948-logs" (OuterVolumeSpecName: "logs") pod "c42e42b6-4389-4938-973d-62cfc12cb948" (UID: "c42e42b6-4389-4938-973d-62cfc12cb948"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.370444 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c42e42b6-4389-4938-973d-62cfc12cb948-logs\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.379289 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c42e42b6-4389-4938-973d-62cfc12cb948-kube-api-access-wztfv" (OuterVolumeSpecName: "kube-api-access-wztfv") pod "c42e42b6-4389-4938-973d-62cfc12cb948" (UID: "c42e42b6-4389-4938-973d-62cfc12cb948"). InnerVolumeSpecName "kube-api-access-wztfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.398157 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c42e42b6-4389-4938-973d-62cfc12cb948" (UID: "c42e42b6-4389-4938-973d-62cfc12cb948"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.408998 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-config-data" (OuterVolumeSpecName: "config-data") pod "c42e42b6-4389-4938-973d-62cfc12cb948" (UID: "c42e42b6-4389-4938-973d-62cfc12cb948"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.430526 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c42e42b6-4389-4938-973d-62cfc12cb948" (UID: "c42e42b6-4389-4938-973d-62cfc12cb948"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.437731 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fc76d242-0072-4bef-a299-0a88624cb734","Type":"ContainerStarted","Data":"5d2c956b2ba53d5263288a2b670fdf6402c9d30476a8424eb4afc94cadd317ed"} Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.437791 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fc76d242-0072-4bef-a299-0a88624cb734","Type":"ContainerStarted","Data":"635493528a58e53f0f9e27e9cb61b0c6b3112d7beea53d8e0d6941f1c783b62e"} Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.442513 4894 generic.go:334] "Generic (PLEG): container finished" podID="c42e42b6-4389-4938-973d-62cfc12cb948" containerID="0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b" exitCode=0 Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.442603 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.442648 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c42e42b6-4389-4938-973d-62cfc12cb948","Type":"ContainerDied","Data":"0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b"} Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.444658 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c42e42b6-4389-4938-973d-62cfc12cb948","Type":"ContainerDied","Data":"bcbdcae84484e1062a65610577466a82e188ab55618599224acdca8f3fca2fcf"} Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.444782 4894 scope.go:117] "RemoveContainer" containerID="0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.463527 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.463502333 podStartE2EDuration="2.463502333s" podCreationTimestamp="2025-12-08 15:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:07:24.457511406 +0000 UTC m=+1265.557517531" watchObservedRunningTime="2025-12-08 15:07:24.463502333 +0000 UTC m=+1265.563508448" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.485947 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wztfv\" (UniqueName: \"kubernetes.io/projected/c42e42b6-4389-4938-973d-62cfc12cb948-kube-api-access-wztfv\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.486098 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.486272 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.486308 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c42e42b6-4389-4938-973d-62cfc12cb948-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.520276 4894 scope.go:117] "RemoveContainer" containerID="d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.551638 4894 scope.go:117] "RemoveContainer" containerID="0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b" Dec 08 15:07:24 crc kubenswrapper[4894]: E1208 15:07:24.554619 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b\": container with ID starting with 0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b not found: ID does not exist" containerID="0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.554713 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b"} err="failed to get container status \"0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b\": rpc error: code = NotFound desc = could not find container \"0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b\": container with ID starting with 0f7d04f84fb093741d7c57dc56f90336466f4805b91b98553ca5014263555c4b not found: ID does not exist" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.554746 4894 scope.go:117] "RemoveContainer" containerID="d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa" Dec 08 15:07:24 crc kubenswrapper[4894]: E1208 15:07:24.555106 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa\": container with ID starting with d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa not found: ID does not exist" containerID="d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.555133 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa"} err="failed to get container status \"d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa\": rpc error: code = NotFound desc = could not find container \"d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa\": container with ID starting with d826f1b7cd6ccc5036e8d25c1f0551e9383f0ead1c0174bd9bab044f9b39a4fa not found: ID does not exist" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.570359 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.580731 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.591238 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:07:24 crc kubenswrapper[4894]: E1208 15:07:24.591666 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-metadata" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.591682 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-metadata" Dec 08 15:07:24 crc kubenswrapper[4894]: E1208 15:07:24.591722 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-log" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.591728 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-log" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.591926 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-log" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.591948 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" containerName="nova-metadata-metadata" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.593871 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.598787 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.598999 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.604716 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.690385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-config-data\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.690447 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-logs\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.690480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.690506 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.690588 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qkbl\" (UniqueName: \"kubernetes.io/projected/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-kube-api-access-7qkbl\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.792674 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qkbl\" (UniqueName: \"kubernetes.io/projected/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-kube-api-access-7qkbl\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.792852 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-config-data\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.792889 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-logs\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.792922 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.792944 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.798551 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-logs\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.803632 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.803765 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-config-data\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.804266 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.826669 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qkbl\" (UniqueName: \"kubernetes.io/projected/5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097-kube-api-access-7qkbl\") pod \"nova-metadata-0\" (UID: \"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097\") " pod="openstack/nova-metadata-0" Dec 08 15:07:24 crc kubenswrapper[4894]: I1208 15:07:24.925711 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 08 15:07:25 crc kubenswrapper[4894]: I1208 15:07:25.210486 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c42e42b6-4389-4938-973d-62cfc12cb948" path="/var/lib/kubelet/pods/c42e42b6-4389-4938-973d-62cfc12cb948/volumes" Dec 08 15:07:25 crc kubenswrapper[4894]: W1208 15:07:25.421751 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dee1e03_ab11_4e3c_b2f2_dd07ca4d4097.slice/crio-97a21f02c5fccad8ac0db95ebaeff7f4b4e99121506b6cbb3fb9523f2c640bc4 WatchSource:0}: Error finding container 97a21f02c5fccad8ac0db95ebaeff7f4b4e99121506b6cbb3fb9523f2c640bc4: Status 404 returned error can't find the container with id 97a21f02c5fccad8ac0db95ebaeff7f4b4e99121506b6cbb3fb9523f2c640bc4 Dec 08 15:07:25 crc kubenswrapper[4894]: I1208 15:07:25.423683 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 08 15:07:25 crc kubenswrapper[4894]: I1208 15:07:25.456887 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097","Type":"ContainerStarted","Data":"97a21f02c5fccad8ac0db95ebaeff7f4b4e99121506b6cbb3fb9523f2c640bc4"} Dec 08 15:07:26 crc kubenswrapper[4894]: I1208 15:07:26.467305 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097","Type":"ContainerStarted","Data":"18e1bcf037c7c93f10a159a42517d83dde8ff17f6197a022355ecd0674340023"} Dec 08 15:07:26 crc kubenswrapper[4894]: I1208 15:07:26.467915 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097","Type":"ContainerStarted","Data":"6b20a6747947e8c1556b557af29b1d77798b923cef64b68c756180cfde230fee"} Dec 08 15:07:26 crc kubenswrapper[4894]: I1208 15:07:26.508380 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.5083572690000002 podStartE2EDuration="2.508357269s" podCreationTimestamp="2025-12-08 15:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:07:26.494422193 +0000 UTC m=+1267.594428308" watchObservedRunningTime="2025-12-08 15:07:26.508357269 +0000 UTC m=+1267.608363384" Dec 08 15:07:28 crc kubenswrapper[4894]: I1208 15:07:28.184933 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 08 15:07:29 crc kubenswrapper[4894]: I1208 15:07:29.925989 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 15:07:29 crc kubenswrapper[4894]: I1208 15:07:29.926051 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 08 15:07:32 crc kubenswrapper[4894]: I1208 15:07:32.145498 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 15:07:32 crc kubenswrapper[4894]: I1208 15:07:32.147146 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 08 15:07:33 crc kubenswrapper[4894]: I1208 15:07:33.158004 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="635376cd-70d9-4108-942a-927f52ce27cb" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 15:07:33 crc kubenswrapper[4894]: I1208 15:07:33.158029 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="635376cd-70d9-4108-942a-927f52ce27cb" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 15:07:33 crc kubenswrapper[4894]: I1208 15:07:33.185212 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 08 15:07:33 crc kubenswrapper[4894]: I1208 15:07:33.212288 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 08 15:07:33 crc kubenswrapper[4894]: I1208 15:07:33.584221 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 08 15:07:34 crc kubenswrapper[4894]: I1208 15:07:34.926511 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 15:07:34 crc kubenswrapper[4894]: I1208 15:07:34.927234 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 08 15:07:35 crc kubenswrapper[4894]: I1208 15:07:35.943105 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 15:07:35 crc kubenswrapper[4894]: I1208 15:07:35.943289 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 08 15:07:41 crc kubenswrapper[4894]: I1208 15:07:41.848658 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 08 15:07:42 crc kubenswrapper[4894]: I1208 15:07:42.154070 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 15:07:42 crc kubenswrapper[4894]: I1208 15:07:42.154447 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 08 15:07:42 crc kubenswrapper[4894]: I1208 15:07:42.154857 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 15:07:42 crc kubenswrapper[4894]: I1208 15:07:42.154901 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 08 15:07:42 crc kubenswrapper[4894]: I1208 15:07:42.160593 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 15:07:42 crc kubenswrapper[4894]: I1208 15:07:42.162511 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 08 15:07:44 crc kubenswrapper[4894]: I1208 15:07:44.933003 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 15:07:44 crc kubenswrapper[4894]: I1208 15:07:44.933287 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 08 15:07:44 crc kubenswrapper[4894]: I1208 15:07:44.943359 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 15:07:45 crc kubenswrapper[4894]: I1208 15:07:45.682225 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 08 15:07:54 crc kubenswrapper[4894]: I1208 15:07:54.117102 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:07:55 crc kubenswrapper[4894]: I1208 15:07:55.776209 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:07:58 crc kubenswrapper[4894]: I1208 15:07:58.606139 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerName="rabbitmq" containerID="cri-o://d1b87c26301a00fc85781bda83dfabda9df7a6356d16a37d056fb9676ba3b413" gracePeriod=604796 Dec 08 15:07:59 crc kubenswrapper[4894]: I1208 15:07:59.907623 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerName="rabbitmq" containerID="cri-o://c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1" gracePeriod=604796 Dec 08 15:08:04 crc kubenswrapper[4894]: I1208 15:08:04.850950 4894 generic.go:334] "Generic (PLEG): container finished" podID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerID="d1b87c26301a00fc85781bda83dfabda9df7a6356d16a37d056fb9676ba3b413" exitCode=0 Dec 08 15:08:04 crc kubenswrapper[4894]: I1208 15:08:04.851018 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d911c29-3d60-4990-b6ec-240d1acebc86","Type":"ContainerDied","Data":"d1b87c26301a00fc85781bda83dfabda9df7a6356d16a37d056fb9676ba3b413"} Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.197583 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.299768 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-tls\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300024 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-server-conf\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300086 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsjbl\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-kube-api-access-tsjbl\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300145 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d911c29-3d60-4990-b6ec-240d1acebc86-pod-info\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300212 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-confd\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300243 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300359 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-plugins-conf\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300423 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-config-data\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300461 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-plugins\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300514 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-erlang-cookie\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.300555 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d911c29-3d60-4990-b6ec-240d1acebc86-erlang-cookie-secret\") pod \"7d911c29-3d60-4990-b6ec-240d1acebc86\" (UID: \"7d911c29-3d60-4990-b6ec-240d1acebc86\") " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.301357 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.301440 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.301526 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.311148 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d911c29-3d60-4990-b6ec-240d1acebc86-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.319727 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7d911c29-3d60-4990-b6ec-240d1acebc86-pod-info" (OuterVolumeSpecName: "pod-info") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.321999 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-kube-api-access-tsjbl" (OuterVolumeSpecName: "kube-api-access-tsjbl") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "kube-api-access-tsjbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.326388 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.360069 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.399637 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-config-data" (OuterVolumeSpecName: "config-data") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403646 4894 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403704 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403721 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403737 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403753 4894 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d911c29-3d60-4990-b6ec-240d1acebc86-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403767 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403781 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsjbl\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-kube-api-access-tsjbl\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403793 4894 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d911c29-3d60-4990-b6ec-240d1acebc86-pod-info\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.403863 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.421105 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-server-conf" (OuterVolumeSpecName: "server-conf") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.451283 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.486778 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7d911c29-3d60-4990-b6ec-240d1acebc86" (UID: "7d911c29-3d60-4990-b6ec-240d1acebc86"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.504902 4894 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d911c29-3d60-4990-b6ec-240d1acebc86-server-conf\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.504931 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d911c29-3d60-4990-b6ec-240d1acebc86-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.504941 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.861983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d911c29-3d60-4990-b6ec-240d1acebc86","Type":"ContainerDied","Data":"5f91cd22f57fc82e1033fe01b97cae861e9b914df86c4a426df19956ed94501e"} Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.862088 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.862641 4894 scope.go:117] "RemoveContainer" containerID="d1b87c26301a00fc85781bda83dfabda9df7a6356d16a37d056fb9676ba3b413" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.889681 4894 scope.go:117] "RemoveContainer" containerID="340018ff56d632188be324a9d2eb44cf1163ca716d83f89365cb7acedd19ccc4" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.898393 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.908636 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.976013 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:08:05 crc kubenswrapper[4894]: E1208 15:08:05.977298 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerName="rabbitmq" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.977321 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerName="rabbitmq" Dec 08 15:08:05 crc kubenswrapper[4894]: E1208 15:08:05.977375 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerName="setup-container" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.977384 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerName="setup-container" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.977736 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" containerName="rabbitmq" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.979933 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.987579 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.988098 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ttv88" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.988261 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.988404 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.988567 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 08 15:08:05 crc kubenswrapper[4894]: I1208 15:08:05.988692 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.005258 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.025714 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.118622 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8447e7e6-4a92-432c-8a5f-48a4cd10675b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.118689 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.118714 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-config-data\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.118747 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.118777 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.118841 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.118863 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.119094 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.119243 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.119720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22kkl\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-kube-api-access-22kkl\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.119831 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8447e7e6-4a92-432c-8a5f-48a4cd10675b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.222733 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22kkl\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-kube-api-access-22kkl\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.222871 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8447e7e6-4a92-432c-8a5f-48a4cd10675b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.222921 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8447e7e6-4a92-432c-8a5f-48a4cd10675b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.222952 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.222977 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-config-data\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.223012 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.223048 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.223095 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.223125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.223163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.223190 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.223987 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.229682 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.229894 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.230291 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-config-data\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.230344 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.231091 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8447e7e6-4a92-432c-8a5f-48a4cd10675b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.234446 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8447e7e6-4a92-432c-8a5f-48a4cd10675b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.234640 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.235457 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8447e7e6-4a92-432c-8a5f-48a4cd10675b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.235492 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.263300 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22kkl\" (UniqueName: \"kubernetes.io/projected/8447e7e6-4a92-432c-8a5f-48a4cd10675b-kube-api-access-22kkl\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.270668 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8447e7e6-4a92-432c-8a5f-48a4cd10675b\") " pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.337141 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.439334 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.531671 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-erlang-cookie\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532275 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-server-conf\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532320 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-plugins-conf\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532335 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532361 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-config-data\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532397 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-plugins\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532457 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbskj\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-kube-api-access-fbskj\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532558 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-erlang-cookie-secret\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532583 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-pod-info\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532761 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-tls\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532788 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.532851 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-confd\") pod \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\" (UID: \"a4dd5876-f3d0-4256-9524-ff3bf39ea28a\") " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.533352 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.538002 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.538726 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.544295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-kube-api-access-fbskj" (OuterVolumeSpecName: "kube-api-access-fbskj") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "kube-api-access-fbskj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.547659 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.553085 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.555354 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.561553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-pod-info" (OuterVolumeSpecName: "pod-info") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.591503 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-config-data" (OuterVolumeSpecName: "config-data") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.618035 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-server-conf" (OuterVolumeSpecName: "server-conf") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.638418 4894 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.638500 4894 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-pod-info\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.640089 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.640128 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.640140 4894 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-server-conf\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.640153 4894 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.640168 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.640182 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.640194 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbskj\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-kube-api-access-fbskj\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.671231 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a4dd5876-f3d0-4256-9524-ff3bf39ea28a" (UID: "a4dd5876-f3d0-4256-9524-ff3bf39ea28a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.674429 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.742669 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.742700 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dd5876-f3d0-4256-9524-ff3bf39ea28a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.877556 4894 generic.go:334] "Generic (PLEG): container finished" podID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerID="c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1" exitCode=0 Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.877615 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a4dd5876-f3d0-4256-9524-ff3bf39ea28a","Type":"ContainerDied","Data":"c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1"} Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.877643 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a4dd5876-f3d0-4256-9524-ff3bf39ea28a","Type":"ContainerDied","Data":"596f1a6693a6a0e395b0773bfbc593ee4ff2afd5ae80e7565c7d0b18d7940f2d"} Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.877660 4894 scope.go:117] "RemoveContainer" containerID="c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.877758 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.910537 4894 scope.go:117] "RemoveContainer" containerID="8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.916680 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.930537 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.943062 4894 scope.go:117] "RemoveContainer" containerID="c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1" Dec 08 15:08:06 crc kubenswrapper[4894]: E1208 15:08:06.943548 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1\": container with ID starting with c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1 not found: ID does not exist" containerID="c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.943606 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1"} err="failed to get container status \"c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1\": rpc error: code = NotFound desc = could not find container \"c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1\": container with ID starting with c1669fda005dd0ab1cabdaa501ab45f7c672a5a165002fbfe19a64a8b6aaaee1 not found: ID does not exist" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.943636 4894 scope.go:117] "RemoveContainer" containerID="8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d" Dec 08 15:08:06 crc kubenswrapper[4894]: E1208 15:08:06.944122 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d\": container with ID starting with 8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d not found: ID does not exist" containerID="8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.944146 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d"} err="failed to get container status \"8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d\": rpc error: code = NotFound desc = could not find container \"8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d\": container with ID starting with 8335f5cf9ed481e3769fb4a9ece401225c1d959363dbd1c93fa39812ca822f7d not found: ID does not exist" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.951907 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:08:06 crc kubenswrapper[4894]: E1208 15:08:06.952613 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerName="rabbitmq" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.952710 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerName="rabbitmq" Dec 08 15:08:06 crc kubenswrapper[4894]: E1208 15:08:06.952805 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerName="setup-container" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.952977 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerName="setup-container" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.953280 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" containerName="rabbitmq" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.955009 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.959180 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.959658 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.960234 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.960465 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.961807 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.962244 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-544nh" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.962583 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.980968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 08 15:08:06 crc kubenswrapper[4894]: I1208 15:08:06.990370 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.057770 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfqtq\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-kube-api-access-dfqtq\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.058156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.058272 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.058404 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba215854-1f7f-4315-acf0-edd51b2358cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.059419 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.059613 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.059955 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.060136 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba215854-1f7f-4315-acf0-edd51b2358cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.060248 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.060388 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.060688 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.162273 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfqtq\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-kube-api-access-dfqtq\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.162693 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.162716 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.163694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.163779 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba215854-1f7f-4315-acf0-edd51b2358cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164358 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164399 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164438 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba215854-1f7f-4315-acf0-edd51b2358cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164545 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164738 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.165538 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.164078 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.166118 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.166893 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba215854-1f7f-4315-acf0-edd51b2358cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.167267 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.169039 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.169626 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.172627 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba215854-1f7f-4315-acf0-edd51b2358cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.176653 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba215854-1f7f-4315-acf0-edd51b2358cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.186295 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfqtq\" (UniqueName: \"kubernetes.io/projected/ba215854-1f7f-4315-acf0-edd51b2358cb-kube-api-access-dfqtq\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.212409 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ba215854-1f7f-4315-acf0-edd51b2358cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.214989 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d911c29-3d60-4990-b6ec-240d1acebc86" path="/var/lib/kubelet/pods/7d911c29-3d60-4990-b6ec-240d1acebc86/volumes" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.216159 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4dd5876-f3d0-4256-9524-ff3bf39ea28a" path="/var/lib/kubelet/pods/a4dd5876-f3d0-4256-9524-ff3bf39ea28a/volumes" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.296658 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.298182 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.477288 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:07 crc kubenswrapper[4894]: I1208 15:08:07.892433 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8447e7e6-4a92-432c-8a5f-48a4cd10675b","Type":"ContainerStarted","Data":"3758a1b420f110b717af72124640f133a820b45b93b6c28bff3710eda76ea925"} Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.013880 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.534473 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-rc4tx"] Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.554787 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.559296 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.579587 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-rc4tx"] Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.592204 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-svc\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.592263 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.592393 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.592439 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nt8r\" (UniqueName: \"kubernetes.io/projected/a8bc04f0-cc08-4004-bb58-b4e9a985857d-kube-api-access-7nt8r\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.592472 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.592519 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-config\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.592543 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nt8r\" (UniqueName: \"kubernetes.io/projected/a8bc04f0-cc08-4004-bb58-b4e9a985857d-kube-api-access-7nt8r\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694079 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694121 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-config\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694139 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694199 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-svc\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694223 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694291 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.694960 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.695096 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.695124 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.695523 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-config\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.695579 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-svc\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.695737 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.713579 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nt8r\" (UniqueName: \"kubernetes.io/projected/a8bc04f0-cc08-4004-bb58-b4e9a985857d-kube-api-access-7nt8r\") pod \"dnsmasq-dns-67b789f86c-rc4tx\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.889721 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.902871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ba215854-1f7f-4315-acf0-edd51b2358cb","Type":"ContainerStarted","Data":"beb9b92c09d43c524f6454810cb64b31051d7eaca58b4d3fd31c584930024af6"} Dec 08 15:08:08 crc kubenswrapper[4894]: I1208 15:08:08.909006 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8447e7e6-4a92-432c-8a5f-48a4cd10675b","Type":"ContainerStarted","Data":"3abed40fef80bee3666f1193a9d45f2de99e5948011b8e652cdbe369f037c1da"} Dec 08 15:08:09 crc kubenswrapper[4894]: I1208 15:08:09.362435 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-rc4tx"] Dec 08 15:08:09 crc kubenswrapper[4894]: W1208 15:08:09.369828 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8bc04f0_cc08_4004_bb58_b4e9a985857d.slice/crio-6c4f7d62d0c9b0a108eb0907e25ff9ed6c6321c75412b62d50106f17ce7d42d7 WatchSource:0}: Error finding container 6c4f7d62d0c9b0a108eb0907e25ff9ed6c6321c75412b62d50106f17ce7d42d7: Status 404 returned error can't find the container with id 6c4f7d62d0c9b0a108eb0907e25ff9ed6c6321c75412b62d50106f17ce7d42d7 Dec 08 15:08:09 crc kubenswrapper[4894]: I1208 15:08:09.919451 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" event={"ID":"a8bc04f0-cc08-4004-bb58-b4e9a985857d","Type":"ContainerStarted","Data":"6c4f7d62d0c9b0a108eb0907e25ff9ed6c6321c75412b62d50106f17ce7d42d7"} Dec 08 15:08:10 crc kubenswrapper[4894]: I1208 15:08:10.929965 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ba215854-1f7f-4315-acf0-edd51b2358cb","Type":"ContainerStarted","Data":"3c2a5e3a807cd24bb0c3d72a36156c90724f125c5e89bdc8ab24722dfaca7cbd"} Dec 08 15:08:10 crc kubenswrapper[4894]: I1208 15:08:10.931598 4894 generic.go:334] "Generic (PLEG): container finished" podID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerID="71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666" exitCode=0 Dec 08 15:08:10 crc kubenswrapper[4894]: I1208 15:08:10.931645 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" event={"ID":"a8bc04f0-cc08-4004-bb58-b4e9a985857d","Type":"ContainerDied","Data":"71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666"} Dec 08 15:08:11 crc kubenswrapper[4894]: I1208 15:08:11.941575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" event={"ID":"a8bc04f0-cc08-4004-bb58-b4e9a985857d","Type":"ContainerStarted","Data":"c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a"} Dec 08 15:08:11 crc kubenswrapper[4894]: I1208 15:08:11.968860 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" podStartSLOduration=3.9688401129999997 podStartE2EDuration="3.968840113s" podCreationTimestamp="2025-12-08 15:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:08:11.959104979 +0000 UTC m=+1313.059111094" watchObservedRunningTime="2025-12-08 15:08:11.968840113 +0000 UTC m=+1313.068846228" Dec 08 15:08:12 crc kubenswrapper[4894]: I1208 15:08:12.951840 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:18 crc kubenswrapper[4894]: I1208 15:08:18.892103 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:18 crc kubenswrapper[4894]: I1208 15:08:18.954302 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-s5t45"] Dec 08 15:08:18 crc kubenswrapper[4894]: I1208 15:08:18.954828 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" podUID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerName="dnsmasq-dns" containerID="cri-o://3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653" gracePeriod=10 Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.102469 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-nb62g"] Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.104907 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.110933 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.110987 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q57n\" (UniqueName: \"kubernetes.io/projected/487bbd0b-1201-4a53-b475-86f85f5a7f51-kube-api-access-4q57n\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.111097 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.111211 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.111304 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.111500 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-config\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.115979 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.149184 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-nb62g"] Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.217539 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-config\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.217947 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.218122 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.218251 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q57n\" (UniqueName: \"kubernetes.io/projected/487bbd0b-1201-4a53-b475-86f85f5a7f51-kube-api-access-4q57n\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.218373 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.218484 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.218595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.225724 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.226530 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.227190 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.229633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.229851 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-config\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.230899 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/487bbd0b-1201-4a53-b475-86f85f5a7f51-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.272318 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q57n\" (UniqueName: \"kubernetes.io/projected/487bbd0b-1201-4a53-b475-86f85f5a7f51-kube-api-access-4q57n\") pod \"dnsmasq-dns-cb6ffcf87-nb62g\" (UID: \"487bbd0b-1201-4a53-b475-86f85f5a7f51\") " pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.468598 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.630859 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.738899 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-svc\") pod \"de234281-8cd3-4b19-8cab-1eff756ff40c\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.738971 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-config\") pod \"de234281-8cd3-4b19-8cab-1eff756ff40c\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.739041 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-sb\") pod \"de234281-8cd3-4b19-8cab-1eff756ff40c\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.739093 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-swift-storage-0\") pod \"de234281-8cd3-4b19-8cab-1eff756ff40c\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.739176 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-nb\") pod \"de234281-8cd3-4b19-8cab-1eff756ff40c\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.739279 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjvpm\" (UniqueName: \"kubernetes.io/projected/de234281-8cd3-4b19-8cab-1eff756ff40c-kube-api-access-jjvpm\") pod \"de234281-8cd3-4b19-8cab-1eff756ff40c\" (UID: \"de234281-8cd3-4b19-8cab-1eff756ff40c\") " Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.752516 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de234281-8cd3-4b19-8cab-1eff756ff40c-kube-api-access-jjvpm" (OuterVolumeSpecName: "kube-api-access-jjvpm") pod "de234281-8cd3-4b19-8cab-1eff756ff40c" (UID: "de234281-8cd3-4b19-8cab-1eff756ff40c"). InnerVolumeSpecName "kube-api-access-jjvpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.812358 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-nb62g"] Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.818994 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "de234281-8cd3-4b19-8cab-1eff756ff40c" (UID: "de234281-8cd3-4b19-8cab-1eff756ff40c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.828196 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-config" (OuterVolumeSpecName: "config") pod "de234281-8cd3-4b19-8cab-1eff756ff40c" (UID: "de234281-8cd3-4b19-8cab-1eff756ff40c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.828388 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de234281-8cd3-4b19-8cab-1eff756ff40c" (UID: "de234281-8cd3-4b19-8cab-1eff756ff40c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.831998 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de234281-8cd3-4b19-8cab-1eff756ff40c" (UID: "de234281-8cd3-4b19-8cab-1eff756ff40c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.835072 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de234281-8cd3-4b19-8cab-1eff756ff40c" (UID: "de234281-8cd3-4b19-8cab-1eff756ff40c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.841606 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.841653 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjvpm\" (UniqueName: \"kubernetes.io/projected/de234281-8cd3-4b19-8cab-1eff756ff40c-kube-api-access-jjvpm\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.841667 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.841678 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.841687 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:19 crc kubenswrapper[4894]: I1208 15:08:19.841695 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de234281-8cd3-4b19-8cab-1eff756ff40c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.021166 4894 generic.go:334] "Generic (PLEG): container finished" podID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerID="3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653" exitCode=0 Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.021351 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.021876 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" event={"ID":"de234281-8cd3-4b19-8cab-1eff756ff40c","Type":"ContainerDied","Data":"3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653"} Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.021930 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-s5t45" event={"ID":"de234281-8cd3-4b19-8cab-1eff756ff40c","Type":"ContainerDied","Data":"4a139b2934c5fe28396f15e0147bbe2a26ef511c1b11dad83c247b71614d8e89"} Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.021951 4894 scope.go:117] "RemoveContainer" containerID="3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653" Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.026939 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" event={"ID":"487bbd0b-1201-4a53-b475-86f85f5a7f51","Type":"ContainerStarted","Data":"639f1b98d02a3b5dbfd74f1d4c20351d76c7ab122477538ee279204532f78a7d"} Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.052969 4894 scope.go:117] "RemoveContainer" containerID="9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426" Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.073776 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-s5t45"] Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.083448 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-s5t45"] Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.088969 4894 scope.go:117] "RemoveContainer" containerID="3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653" Dec 08 15:08:20 crc kubenswrapper[4894]: E1208 15:08:20.089641 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653\": container with ID starting with 3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653 not found: ID does not exist" containerID="3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653" Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.089764 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653"} err="failed to get container status \"3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653\": rpc error: code = NotFound desc = could not find container \"3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653\": container with ID starting with 3b52f7cb638d5c81172140f0eef8eb2a105df7fc587aa358be9719e5a6acd653 not found: ID does not exist" Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.089895 4894 scope.go:117] "RemoveContainer" containerID="9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426" Dec 08 15:08:20 crc kubenswrapper[4894]: E1208 15:08:20.090361 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426\": container with ID starting with 9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426 not found: ID does not exist" containerID="9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426" Dec 08 15:08:20 crc kubenswrapper[4894]: I1208 15:08:20.090477 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426"} err="failed to get container status \"9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426\": rpc error: code = NotFound desc = could not find container \"9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426\": container with ID starting with 9a35be7c3fd2c88908c620e763368170e60f8a27d8ae15076ea6b7f834084426 not found: ID does not exist" Dec 08 15:08:21 crc kubenswrapper[4894]: I1208 15:08:21.037642 4894 generic.go:334] "Generic (PLEG): container finished" podID="487bbd0b-1201-4a53-b475-86f85f5a7f51" containerID="33d287f39b85fd07c519c67aead48f6d35cf98f4f4c023691baf14c87167335e" exitCode=0 Dec 08 15:08:21 crc kubenswrapper[4894]: I1208 15:08:21.037731 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" event={"ID":"487bbd0b-1201-4a53-b475-86f85f5a7f51","Type":"ContainerDied","Data":"33d287f39b85fd07c519c67aead48f6d35cf98f4f4c023691baf14c87167335e"} Dec 08 15:08:21 crc kubenswrapper[4894]: I1208 15:08:21.246074 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de234281-8cd3-4b19-8cab-1eff756ff40c" path="/var/lib/kubelet/pods/de234281-8cd3-4b19-8cab-1eff756ff40c/volumes" Dec 08 15:08:22 crc kubenswrapper[4894]: I1208 15:08:22.059525 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" event={"ID":"487bbd0b-1201-4a53-b475-86f85f5a7f51","Type":"ContainerStarted","Data":"3638e5af9f2d056e4cead20ee68427603a2474797f01d257ffef22f18e2f60a9"} Dec 08 15:08:22 crc kubenswrapper[4894]: I1208 15:08:22.060931 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:22 crc kubenswrapper[4894]: I1208 15:08:22.081482 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" podStartSLOduration=3.081466451 podStartE2EDuration="3.081466451s" podCreationTimestamp="2025-12-08 15:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:08:22.079440578 +0000 UTC m=+1323.179446693" watchObservedRunningTime="2025-12-08 15:08:22.081466451 +0000 UTC m=+1323.181472566" Dec 08 15:08:29 crc kubenswrapper[4894]: I1208 15:08:29.471041 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-nb62g" Dec 08 15:08:29 crc kubenswrapper[4894]: I1208 15:08:29.557977 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-rc4tx"] Dec 08 15:08:29 crc kubenswrapper[4894]: I1208 15:08:29.558221 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" podUID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerName="dnsmasq-dns" containerID="cri-o://c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a" gracePeriod=10 Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.084990 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.149197 4894 generic.go:334] "Generic (PLEG): container finished" podID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerID="c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a" exitCode=0 Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.149231 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" event={"ID":"a8bc04f0-cc08-4004-bb58-b4e9a985857d","Type":"ContainerDied","Data":"c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a"} Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.149256 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" event={"ID":"a8bc04f0-cc08-4004-bb58-b4e9a985857d","Type":"ContainerDied","Data":"6c4f7d62d0c9b0a108eb0907e25ff9ed6c6321c75412b62d50106f17ce7d42d7"} Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.149273 4894 scope.go:117] "RemoveContainer" containerID="c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.149290 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-rc4tx" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.186559 4894 scope.go:117] "RemoveContainer" containerID="71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.220971 4894 scope.go:117] "RemoveContainer" containerID="c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a" Dec 08 15:08:30 crc kubenswrapper[4894]: E1208 15:08:30.222278 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a\": container with ID starting with c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a not found: ID does not exist" containerID="c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.222312 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a"} err="failed to get container status \"c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a\": rpc error: code = NotFound desc = could not find container \"c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a\": container with ID starting with c33452fef7282bc42211d9d93fa41e2b698c24f98497cbcde4c531c8557c4e4a not found: ID does not exist" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.222336 4894 scope.go:117] "RemoveContainer" containerID="71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666" Dec 08 15:08:30 crc kubenswrapper[4894]: E1208 15:08:30.227114 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666\": container with ID starting with 71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666 not found: ID does not exist" containerID="71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.227147 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666"} err="failed to get container status \"71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666\": rpc error: code = NotFound desc = could not find container \"71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666\": container with ID starting with 71a08e429ea191cf03d6821d6846734590ba9519567e3c40e0ed540aad4ca666 not found: ID does not exist" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.239527 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-swift-storage-0\") pod \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.239652 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-sb\") pod \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.239760 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-nb\") pod \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.239803 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-svc\") pod \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.239887 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-config\") pod \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.239935 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nt8r\" (UniqueName: \"kubernetes.io/projected/a8bc04f0-cc08-4004-bb58-b4e9a985857d-kube-api-access-7nt8r\") pod \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.239958 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-openstack-edpm-ipam\") pod \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\" (UID: \"a8bc04f0-cc08-4004-bb58-b4e9a985857d\") " Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.313001 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8bc04f0-cc08-4004-bb58-b4e9a985857d-kube-api-access-7nt8r" (OuterVolumeSpecName: "kube-api-access-7nt8r") pod "a8bc04f0-cc08-4004-bb58-b4e9a985857d" (UID: "a8bc04f0-cc08-4004-bb58-b4e9a985857d"). InnerVolumeSpecName "kube-api-access-7nt8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.342159 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nt8r\" (UniqueName: \"kubernetes.io/projected/a8bc04f0-cc08-4004-bb58-b4e9a985857d-kube-api-access-7nt8r\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.401418 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a8bc04f0-cc08-4004-bb58-b4e9a985857d" (UID: "a8bc04f0-cc08-4004-bb58-b4e9a985857d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.444354 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.456423 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-config" (OuterVolumeSpecName: "config") pod "a8bc04f0-cc08-4004-bb58-b4e9a985857d" (UID: "a8bc04f0-cc08-4004-bb58-b4e9a985857d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.467145 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a8bc04f0-cc08-4004-bb58-b4e9a985857d" (UID: "a8bc04f0-cc08-4004-bb58-b4e9a985857d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.477131 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a8bc04f0-cc08-4004-bb58-b4e9a985857d" (UID: "a8bc04f0-cc08-4004-bb58-b4e9a985857d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.491602 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a8bc04f0-cc08-4004-bb58-b4e9a985857d" (UID: "a8bc04f0-cc08-4004-bb58-b4e9a985857d"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.498089 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a8bc04f0-cc08-4004-bb58-b4e9a985857d" (UID: "a8bc04f0-cc08-4004-bb58-b4e9a985857d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.547228 4894 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.547269 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.547282 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.547294 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.547306 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a8bc04f0-cc08-4004-bb58-b4e9a985857d-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.782790 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-rc4tx"] Dec 08 15:08:30 crc kubenswrapper[4894]: I1208 15:08:30.791911 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-rc4tx"] Dec 08 15:08:31 crc kubenswrapper[4894]: I1208 15:08:31.223298 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" path="/var/lib/kubelet/pods/a8bc04f0-cc08-4004-bb58-b4e9a985857d/volumes" Dec 08 15:08:37 crc kubenswrapper[4894]: I1208 15:08:37.296884 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:08:37 crc kubenswrapper[4894]: I1208 15:08:37.297542 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:08:41 crc kubenswrapper[4894]: I1208 15:08:41.245181 4894 generic.go:334] "Generic (PLEG): container finished" podID="8447e7e6-4a92-432c-8a5f-48a4cd10675b" containerID="3abed40fef80bee3666f1193a9d45f2de99e5948011b8e652cdbe369f037c1da" exitCode=0 Dec 08 15:08:41 crc kubenswrapper[4894]: I1208 15:08:41.245225 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8447e7e6-4a92-432c-8a5f-48a4cd10675b","Type":"ContainerDied","Data":"3abed40fef80bee3666f1193a9d45f2de99e5948011b8e652cdbe369f037c1da"} Dec 08 15:08:42 crc kubenswrapper[4894]: I1208 15:08:42.258396 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8447e7e6-4a92-432c-8a5f-48a4cd10675b","Type":"ContainerStarted","Data":"50f2e769899c0f3a0641d5f1d9a38eb39803df20e412039f12e88a07df109692"} Dec 08 15:08:42 crc kubenswrapper[4894]: I1208 15:08:42.259184 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 08 15:08:42 crc kubenswrapper[4894]: I1208 15:08:42.291601 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.291582565 podStartE2EDuration="37.291582565s" podCreationTimestamp="2025-12-08 15:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:08:42.285176675 +0000 UTC m=+1343.385182800" watchObservedRunningTime="2025-12-08 15:08:42.291582565 +0000 UTC m=+1343.391588680" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.027336 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7"] Dec 08 15:08:43 crc kubenswrapper[4894]: E1208 15:08:43.028051 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerName="dnsmasq-dns" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.028067 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerName="dnsmasq-dns" Dec 08 15:08:43 crc kubenswrapper[4894]: E1208 15:08:43.028082 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerName="dnsmasq-dns" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.028088 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerName="dnsmasq-dns" Dec 08 15:08:43 crc kubenswrapper[4894]: E1208 15:08:43.028096 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerName="init" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.028104 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerName="init" Dec 08 15:08:43 crc kubenswrapper[4894]: E1208 15:08:43.028120 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerName="init" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.028126 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerName="init" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.028289 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="de234281-8cd3-4b19-8cab-1eff756ff40c" containerName="dnsmasq-dns" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.028311 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8bc04f0-cc08-4004-bb58-b4e9a985857d" containerName="dnsmasq-dns" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.028964 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.031660 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.032199 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.039067 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.039654 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.048113 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7"] Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.195886 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.195953 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw7kh\" (UniqueName: \"kubernetes.io/projected/211799eb-bf07-4965-8404-ec3ddba08c3b-kube-api-access-vw7kh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.196033 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.196070 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.298916 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw7kh\" (UniqueName: \"kubernetes.io/projected/211799eb-bf07-4965-8404-ec3ddba08c3b-kube-api-access-vw7kh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.299059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.299115 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.299355 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.310721 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.319499 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.320660 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.335215 4894 generic.go:334] "Generic (PLEG): container finished" podID="ba215854-1f7f-4315-acf0-edd51b2358cb" containerID="3c2a5e3a807cd24bb0c3d72a36156c90724f125c5e89bdc8ab24722dfaca7cbd" exitCode=0 Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.335560 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ba215854-1f7f-4315-acf0-edd51b2358cb","Type":"ContainerDied","Data":"3c2a5e3a807cd24bb0c3d72a36156c90724f125c5e89bdc8ab24722dfaca7cbd"} Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.338519 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw7kh\" (UniqueName: \"kubernetes.io/projected/211799eb-bf07-4965-8404-ec3ddba08c3b-kube-api-access-vw7kh\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.390240 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:08:43 crc kubenswrapper[4894]: I1208 15:08:43.960893 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7"] Dec 08 15:08:44 crc kubenswrapper[4894]: I1208 15:08:44.345667 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" event={"ID":"211799eb-bf07-4965-8404-ec3ddba08c3b","Type":"ContainerStarted","Data":"5b9f25a8b4ff96b0a7e01efd2de907b13a10e0f4e387f2e672a7f26fdc806f56"} Dec 08 15:08:44 crc kubenswrapper[4894]: I1208 15:08:44.348195 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ba215854-1f7f-4315-acf0-edd51b2358cb","Type":"ContainerStarted","Data":"2b16949e4e93e03205fb7e6912961a2d2e8ed08f67ff8a6aabb491d5a38f3a20"} Dec 08 15:08:44 crc kubenswrapper[4894]: I1208 15:08:44.348498 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:08:44 crc kubenswrapper[4894]: I1208 15:08:44.374194 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.37417733 podStartE2EDuration="38.37417733s" podCreationTimestamp="2025-12-08 15:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:08:44.370410692 +0000 UTC m=+1345.470416797" watchObservedRunningTime="2025-12-08 15:08:44.37417733 +0000 UTC m=+1345.474183445" Dec 08 15:08:53 crc kubenswrapper[4894]: I1208 15:08:53.447268 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" event={"ID":"211799eb-bf07-4965-8404-ec3ddba08c3b","Type":"ContainerStarted","Data":"4007c756467cd01844531bb8e2b4e3828aa14cbaacb0d7981a0a3e6126e8815e"} Dec 08 15:08:53 crc kubenswrapper[4894]: I1208 15:08:53.473763 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" podStartSLOduration=2.429169768 podStartE2EDuration="11.473744218s" podCreationTimestamp="2025-12-08 15:08:42 +0000 UTC" firstStartedPulling="2025-12-08 15:08:43.968867817 +0000 UTC m=+1345.068873932" lastFinishedPulling="2025-12-08 15:08:53.013442257 +0000 UTC m=+1354.113448382" observedRunningTime="2025-12-08 15:08:53.463609141 +0000 UTC m=+1354.563615256" watchObservedRunningTime="2025-12-08 15:08:53.473744218 +0000 UTC m=+1354.573750333" Dec 08 15:08:56 crc kubenswrapper[4894]: I1208 15:08:56.342117 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 08 15:08:57 crc kubenswrapper[4894]: I1208 15:08:57.486042 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 08 15:09:04 crc kubenswrapper[4894]: I1208 15:09:04.558108 4894 generic.go:334] "Generic (PLEG): container finished" podID="211799eb-bf07-4965-8404-ec3ddba08c3b" containerID="4007c756467cd01844531bb8e2b4e3828aa14cbaacb0d7981a0a3e6126e8815e" exitCode=0 Dec 08 15:09:04 crc kubenswrapper[4894]: I1208 15:09:04.558191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" event={"ID":"211799eb-bf07-4965-8404-ec3ddba08c3b","Type":"ContainerDied","Data":"4007c756467cd01844531bb8e2b4e3828aa14cbaacb0d7981a0a3e6126e8815e"} Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.111050 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.141296 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-ssh-key\") pod \"211799eb-bf07-4965-8404-ec3ddba08c3b\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.141438 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-repo-setup-combined-ca-bundle\") pod \"211799eb-bf07-4965-8404-ec3ddba08c3b\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.141542 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw7kh\" (UniqueName: \"kubernetes.io/projected/211799eb-bf07-4965-8404-ec3ddba08c3b-kube-api-access-vw7kh\") pod \"211799eb-bf07-4965-8404-ec3ddba08c3b\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.141575 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-inventory\") pod \"211799eb-bf07-4965-8404-ec3ddba08c3b\" (UID: \"211799eb-bf07-4965-8404-ec3ddba08c3b\") " Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.151285 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/211799eb-bf07-4965-8404-ec3ddba08c3b-kube-api-access-vw7kh" (OuterVolumeSpecName: "kube-api-access-vw7kh") pod "211799eb-bf07-4965-8404-ec3ddba08c3b" (UID: "211799eb-bf07-4965-8404-ec3ddba08c3b"). InnerVolumeSpecName "kube-api-access-vw7kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.151247 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "211799eb-bf07-4965-8404-ec3ddba08c3b" (UID: "211799eb-bf07-4965-8404-ec3ddba08c3b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.224097 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "211799eb-bf07-4965-8404-ec3ddba08c3b" (UID: "211799eb-bf07-4965-8404-ec3ddba08c3b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.227083 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-inventory" (OuterVolumeSpecName: "inventory") pod "211799eb-bf07-4965-8404-ec3ddba08c3b" (UID: "211799eb-bf07-4965-8404-ec3ddba08c3b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.260457 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.260505 4894 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.260531 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw7kh\" (UniqueName: \"kubernetes.io/projected/211799eb-bf07-4965-8404-ec3ddba08c3b-kube-api-access-vw7kh\") on node \"crc\" DevicePath \"\"" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.260549 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/211799eb-bf07-4965-8404-ec3ddba08c3b-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.578139 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" event={"ID":"211799eb-bf07-4965-8404-ec3ddba08c3b","Type":"ContainerDied","Data":"5b9f25a8b4ff96b0a7e01efd2de907b13a10e0f4e387f2e672a7f26fdc806f56"} Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.578172 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.578185 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b9f25a8b4ff96b0a7e01efd2de907b13a10e0f4e387f2e672a7f26fdc806f56" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.656761 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq"] Dec 08 15:09:06 crc kubenswrapper[4894]: E1208 15:09:06.657206 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="211799eb-bf07-4965-8404-ec3ddba08c3b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.657222 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="211799eb-bf07-4965-8404-ec3ddba08c3b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.657414 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="211799eb-bf07-4965-8404-ec3ddba08c3b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.658211 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.660504 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.660729 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.661128 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.665239 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.666690 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq"] Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.769065 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs9v4\" (UniqueName: \"kubernetes.io/projected/0147e712-181c-4b8e-85d3-3b90c985f5ee-kube-api-access-bs9v4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.769195 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.769228 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.871375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs9v4\" (UniqueName: \"kubernetes.io/projected/0147e712-181c-4b8e-85d3-3b90c985f5ee-kube-api-access-bs9v4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.871521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.871563 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.875571 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.875640 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.885668 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs9v4\" (UniqueName: \"kubernetes.io/projected/0147e712-181c-4b8e-85d3-3b90c985f5ee-kube-api-access-bs9v4\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qsrmq\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:06 crc kubenswrapper[4894]: I1208 15:09:06.976296 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.297219 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.297572 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.297624 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.298423 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9103dd6f195ccf766c15c1a9fa81340683956ac05b3d32c6e73e01c28e7fa11"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.298476 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://d9103dd6f195ccf766c15c1a9fa81340683956ac05b3d32c6e73e01c28e7fa11" gracePeriod=600 Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.596001 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="d9103dd6f195ccf766c15c1a9fa81340683956ac05b3d32c6e73e01c28e7fa11" exitCode=0 Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.596557 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"d9103dd6f195ccf766c15c1a9fa81340683956ac05b3d32c6e73e01c28e7fa11"} Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.596659 4894 scope.go:117] "RemoveContainer" containerID="9002401f278cd2ddc27524b106a2357f8f007e534b8798b7fd2145f577807600" Dec 08 15:09:07 crc kubenswrapper[4894]: I1208 15:09:07.695833 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq"] Dec 08 15:09:08 crc kubenswrapper[4894]: I1208 15:09:08.607915 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" event={"ID":"0147e712-181c-4b8e-85d3-3b90c985f5ee","Type":"ContainerStarted","Data":"3f3e8e9020eb5a7061cbbcac60bb1066dc71bfe66305a68ef7278ff275298a41"} Dec 08 15:09:08 crc kubenswrapper[4894]: I1208 15:09:08.608512 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" event={"ID":"0147e712-181c-4b8e-85d3-3b90c985f5ee","Type":"ContainerStarted","Data":"632fa4a843b844193b2d590101e14eb7563693f5d1a9a79d9322c4dfd1e9628f"} Dec 08 15:09:08 crc kubenswrapper[4894]: I1208 15:09:08.610725 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa"} Dec 08 15:09:08 crc kubenswrapper[4894]: I1208 15:09:08.623970 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" podStartSLOduration=2.166136958 podStartE2EDuration="2.623951546s" podCreationTimestamp="2025-12-08 15:09:06 +0000 UTC" firstStartedPulling="2025-12-08 15:09:07.705329979 +0000 UTC m=+1368.805336094" lastFinishedPulling="2025-12-08 15:09:08.163144567 +0000 UTC m=+1369.263150682" observedRunningTime="2025-12-08 15:09:08.620890141 +0000 UTC m=+1369.720896256" watchObservedRunningTime="2025-12-08 15:09:08.623951546 +0000 UTC m=+1369.723957661" Dec 08 15:09:11 crc kubenswrapper[4894]: I1208 15:09:11.634841 4894 generic.go:334] "Generic (PLEG): container finished" podID="0147e712-181c-4b8e-85d3-3b90c985f5ee" containerID="3f3e8e9020eb5a7061cbbcac60bb1066dc71bfe66305a68ef7278ff275298a41" exitCode=0 Dec 08 15:09:11 crc kubenswrapper[4894]: I1208 15:09:11.634936 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" event={"ID":"0147e712-181c-4b8e-85d3-3b90c985f5ee","Type":"ContainerDied","Data":"3f3e8e9020eb5a7061cbbcac60bb1066dc71bfe66305a68ef7278ff275298a41"} Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.089091 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.211203 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs9v4\" (UniqueName: \"kubernetes.io/projected/0147e712-181c-4b8e-85d3-3b90c985f5ee-kube-api-access-bs9v4\") pod \"0147e712-181c-4b8e-85d3-3b90c985f5ee\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.211299 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-ssh-key\") pod \"0147e712-181c-4b8e-85d3-3b90c985f5ee\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.211428 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-inventory\") pod \"0147e712-181c-4b8e-85d3-3b90c985f5ee\" (UID: \"0147e712-181c-4b8e-85d3-3b90c985f5ee\") " Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.224031 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0147e712-181c-4b8e-85d3-3b90c985f5ee-kube-api-access-bs9v4" (OuterVolumeSpecName: "kube-api-access-bs9v4") pod "0147e712-181c-4b8e-85d3-3b90c985f5ee" (UID: "0147e712-181c-4b8e-85d3-3b90c985f5ee"). InnerVolumeSpecName "kube-api-access-bs9v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.242971 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-inventory" (OuterVolumeSpecName: "inventory") pod "0147e712-181c-4b8e-85d3-3b90c985f5ee" (UID: "0147e712-181c-4b8e-85d3-3b90c985f5ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.257682 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0147e712-181c-4b8e-85d3-3b90c985f5ee" (UID: "0147e712-181c-4b8e-85d3-3b90c985f5ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.313252 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.313283 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs9v4\" (UniqueName: \"kubernetes.io/projected/0147e712-181c-4b8e-85d3-3b90c985f5ee-kube-api-access-bs9v4\") on node \"crc\" DevicePath \"\"" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.313296 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0147e712-181c-4b8e-85d3-3b90c985f5ee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.655594 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" event={"ID":"0147e712-181c-4b8e-85d3-3b90c985f5ee","Type":"ContainerDied","Data":"632fa4a843b844193b2d590101e14eb7563693f5d1a9a79d9322c4dfd1e9628f"} Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.655998 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="632fa4a843b844193b2d590101e14eb7563693f5d1a9a79d9322c4dfd1e9628f" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.655673 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qsrmq" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.732425 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph"] Dec 08 15:09:13 crc kubenswrapper[4894]: E1208 15:09:13.732920 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0147e712-181c-4b8e-85d3-3b90c985f5ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.732941 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0147e712-181c-4b8e-85d3-3b90c985f5ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.733199 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0147e712-181c-4b8e-85d3-3b90c985f5ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.733997 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.741311 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.741328 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.741466 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.741549 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.746496 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph"] Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.926850 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.926906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.927151 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:13 crc kubenswrapper[4894]: I1208 15:09:13.927326 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxfc2\" (UniqueName: \"kubernetes.io/projected/86e41fca-6958-4b8a-84c0-851b75f8efea-kube-api-access-kxfc2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.028568 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.028644 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxfc2\" (UniqueName: \"kubernetes.io/projected/86e41fca-6958-4b8a-84c0-851b75f8efea-kube-api-access-kxfc2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.028708 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.028738 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.033262 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.033355 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.033582 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.047316 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxfc2\" (UniqueName: \"kubernetes.io/projected/86e41fca-6958-4b8a-84c0-851b75f8efea-kube-api-access-kxfc2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.057438 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.570323 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph"] Dec 08 15:09:14 crc kubenswrapper[4894]: I1208 15:09:14.665508 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" event={"ID":"86e41fca-6958-4b8a-84c0-851b75f8efea","Type":"ContainerStarted","Data":"4bd80ab75e379aaa1e3824e913350b9f0bb5ea9ea011d4781f0ebb08c608bf27"} Dec 08 15:09:15 crc kubenswrapper[4894]: I1208 15:09:15.676019 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" event={"ID":"86e41fca-6958-4b8a-84c0-851b75f8efea","Type":"ContainerStarted","Data":"8f41302a2a3dad69a1baf29bd8dec83d3563c78bace70935536ab3e9888a5c3f"} Dec 08 15:09:15 crc kubenswrapper[4894]: I1208 15:09:15.705498 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" podStartSLOduration=2.33116391 podStartE2EDuration="2.705471642s" podCreationTimestamp="2025-12-08 15:09:13 +0000 UTC" firstStartedPulling="2025-12-08 15:09:14.565996448 +0000 UTC m=+1375.666002563" lastFinishedPulling="2025-12-08 15:09:14.94030418 +0000 UTC m=+1376.040310295" observedRunningTime="2025-12-08 15:09:15.69034412 +0000 UTC m=+1376.790350255" watchObservedRunningTime="2025-12-08 15:09:15.705471642 +0000 UTC m=+1376.805477767" Dec 08 15:09:53 crc kubenswrapper[4894]: I1208 15:09:53.009698 4894 scope.go:117] "RemoveContainer" containerID="19bcc9c3e34208a8586fb01e1a19a1ed4f3c02766ce286e57566ebba290aae9a" Dec 08 15:10:42 crc kubenswrapper[4894]: I1208 15:10:42.876288 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c8sh5"] Dec 08 15:10:42 crc kubenswrapper[4894]: I1208 15:10:42.883644 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:42 crc kubenswrapper[4894]: I1208 15:10:42.890459 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c8sh5"] Dec 08 15:10:42 crc kubenswrapper[4894]: I1208 15:10:42.925788 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-779c6\" (UniqueName: \"kubernetes.io/projected/a232a29c-ae10-425c-abda-7aff71396bcc-kube-api-access-779c6\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:42 crc kubenswrapper[4894]: I1208 15:10:42.926176 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-utilities\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:42 crc kubenswrapper[4894]: I1208 15:10:42.926201 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-catalog-content\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.028204 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-779c6\" (UniqueName: \"kubernetes.io/projected/a232a29c-ae10-425c-abda-7aff71396bcc-kube-api-access-779c6\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.028585 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-utilities\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.028695 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-catalog-content\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.029212 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-utilities\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.029225 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-catalog-content\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.047882 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-779c6\" (UniqueName: \"kubernetes.io/projected/a232a29c-ae10-425c-abda-7aff71396bcc-kube-api-access-779c6\") pod \"community-operators-c8sh5\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.207588 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:43 crc kubenswrapper[4894]: I1208 15:10:43.732365 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c8sh5"] Dec 08 15:10:44 crc kubenswrapper[4894]: I1208 15:10:44.534211 4894 generic.go:334] "Generic (PLEG): container finished" podID="a232a29c-ae10-425c-abda-7aff71396bcc" containerID="4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad" exitCode=0 Dec 08 15:10:44 crc kubenswrapper[4894]: I1208 15:10:44.534266 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8sh5" event={"ID":"a232a29c-ae10-425c-abda-7aff71396bcc","Type":"ContainerDied","Data":"4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad"} Dec 08 15:10:44 crc kubenswrapper[4894]: I1208 15:10:44.534482 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8sh5" event={"ID":"a232a29c-ae10-425c-abda-7aff71396bcc","Type":"ContainerStarted","Data":"182ff0c51da9344ae479e8b8f21e89d24f7c785133e8dc585786ba82906a619d"} Dec 08 15:10:45 crc kubenswrapper[4894]: I1208 15:10:45.554041 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8sh5" event={"ID":"a232a29c-ae10-425c-abda-7aff71396bcc","Type":"ContainerStarted","Data":"fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6"} Dec 08 15:10:46 crc kubenswrapper[4894]: I1208 15:10:46.565387 4894 generic.go:334] "Generic (PLEG): container finished" podID="a232a29c-ae10-425c-abda-7aff71396bcc" containerID="fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6" exitCode=0 Dec 08 15:10:46 crc kubenswrapper[4894]: I1208 15:10:46.565433 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8sh5" event={"ID":"a232a29c-ae10-425c-abda-7aff71396bcc","Type":"ContainerDied","Data":"fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6"} Dec 08 15:10:47 crc kubenswrapper[4894]: I1208 15:10:47.577464 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8sh5" event={"ID":"a232a29c-ae10-425c-abda-7aff71396bcc","Type":"ContainerStarted","Data":"1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce"} Dec 08 15:10:47 crc kubenswrapper[4894]: I1208 15:10:47.598315 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c8sh5" podStartSLOduration=3.142023558 podStartE2EDuration="5.598297155s" podCreationTimestamp="2025-12-08 15:10:42 +0000 UTC" firstStartedPulling="2025-12-08 15:10:44.536133263 +0000 UTC m=+1465.636139378" lastFinishedPulling="2025-12-08 15:10:46.99240686 +0000 UTC m=+1468.092412975" observedRunningTime="2025-12-08 15:10:47.592731891 +0000 UTC m=+1468.692738006" watchObservedRunningTime="2025-12-08 15:10:47.598297155 +0000 UTC m=+1468.698303270" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.084493 4894 scope.go:117] "RemoveContainer" containerID="c594e0c5dcc16df2335f1449b117bfd30533198a74973942f452ee355b0662f7" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.111805 4894 scope.go:117] "RemoveContainer" containerID="548607c0f720255da6e0f5953d7edf31caf09aea918dc6d6c6ddd0e5569d351a" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.149139 4894 scope.go:117] "RemoveContainer" containerID="9633f9c2c01770279de3d05e8a175ad057c9698ed75b2b596ae63ed349607cd4" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.211375 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.211752 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.258494 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.681070 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:53 crc kubenswrapper[4894]: I1208 15:10:53.730952 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c8sh5"] Dec 08 15:10:55 crc kubenswrapper[4894]: I1208 15:10:55.656383 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c8sh5" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="registry-server" containerID="cri-o://1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce" gracePeriod=2 Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.078232 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.189886 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-utilities\") pod \"a232a29c-ae10-425c-abda-7aff71396bcc\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.190009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-779c6\" (UniqueName: \"kubernetes.io/projected/a232a29c-ae10-425c-abda-7aff71396bcc-kube-api-access-779c6\") pod \"a232a29c-ae10-425c-abda-7aff71396bcc\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.190109 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-catalog-content\") pod \"a232a29c-ae10-425c-abda-7aff71396bcc\" (UID: \"a232a29c-ae10-425c-abda-7aff71396bcc\") " Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.191853 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-utilities" (OuterVolumeSpecName: "utilities") pod "a232a29c-ae10-425c-abda-7aff71396bcc" (UID: "a232a29c-ae10-425c-abda-7aff71396bcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.202209 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a232a29c-ae10-425c-abda-7aff71396bcc-kube-api-access-779c6" (OuterVolumeSpecName: "kube-api-access-779c6") pod "a232a29c-ae10-425c-abda-7aff71396bcc" (UID: "a232a29c-ae10-425c-abda-7aff71396bcc"). InnerVolumeSpecName "kube-api-access-779c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.248995 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a232a29c-ae10-425c-abda-7aff71396bcc" (UID: "a232a29c-ae10-425c-abda-7aff71396bcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.292995 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.293046 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a232a29c-ae10-425c-abda-7aff71396bcc-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.293064 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-779c6\" (UniqueName: \"kubernetes.io/projected/a232a29c-ae10-425c-abda-7aff71396bcc-kube-api-access-779c6\") on node \"crc\" DevicePath \"\"" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.666961 4894 generic.go:334] "Generic (PLEG): container finished" podID="a232a29c-ae10-425c-abda-7aff71396bcc" containerID="1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce" exitCode=0 Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.667002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8sh5" event={"ID":"a232a29c-ae10-425c-abda-7aff71396bcc","Type":"ContainerDied","Data":"1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce"} Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.667034 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8sh5" event={"ID":"a232a29c-ae10-425c-abda-7aff71396bcc","Type":"ContainerDied","Data":"182ff0c51da9344ae479e8b8f21e89d24f7c785133e8dc585786ba82906a619d"} Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.667040 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8sh5" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.667051 4894 scope.go:117] "RemoveContainer" containerID="1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.697043 4894 scope.go:117] "RemoveContainer" containerID="fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.703406 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c8sh5"] Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.713321 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c8sh5"] Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.716689 4894 scope.go:117] "RemoveContainer" containerID="4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.775506 4894 scope.go:117] "RemoveContainer" containerID="1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce" Dec 08 15:10:56 crc kubenswrapper[4894]: E1208 15:10:56.776098 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce\": container with ID starting with 1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce not found: ID does not exist" containerID="1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.776149 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce"} err="failed to get container status \"1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce\": rpc error: code = NotFound desc = could not find container \"1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce\": container with ID starting with 1cac303b74db229bbac0920d36d3a7855ab99be8ec2f2357b92f738fa04da9ce not found: ID does not exist" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.776176 4894 scope.go:117] "RemoveContainer" containerID="fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6" Dec 08 15:10:56 crc kubenswrapper[4894]: E1208 15:10:56.776551 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6\": container with ID starting with fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6 not found: ID does not exist" containerID="fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.776598 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6"} err="failed to get container status \"fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6\": rpc error: code = NotFound desc = could not find container \"fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6\": container with ID starting with fb5fc8053a4d79cfc269bd1e85d9f0bf738937b04d6006d02e8c825349a4abb6 not found: ID does not exist" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.776621 4894 scope.go:117] "RemoveContainer" containerID="4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad" Dec 08 15:10:56 crc kubenswrapper[4894]: E1208 15:10:56.777101 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad\": container with ID starting with 4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad not found: ID does not exist" containerID="4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad" Dec 08 15:10:56 crc kubenswrapper[4894]: I1208 15:10:56.777150 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad"} err="failed to get container status \"4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad\": rpc error: code = NotFound desc = could not find container \"4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad\": container with ID starting with 4f3f1ab9483f4cbb3c379cce795bac729a2b8fead990ad6374ed20a491e06cad not found: ID does not exist" Dec 08 15:10:57 crc kubenswrapper[4894]: I1208 15:10:57.206730 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" path="/var/lib/kubelet/pods/a232a29c-ae10-425c-abda-7aff71396bcc/volumes" Dec 08 15:11:07 crc kubenswrapper[4894]: I1208 15:11:07.296686 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:11:07 crc kubenswrapper[4894]: I1208 15:11:07.297643 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:11:37 crc kubenswrapper[4894]: I1208 15:11:37.296926 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:11:37 crc kubenswrapper[4894]: I1208 15:11:37.297457 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:11:53 crc kubenswrapper[4894]: I1208 15:11:53.234623 4894 scope.go:117] "RemoveContainer" containerID="a610c2bb1259b7de3dc83510b80c961c9308e95254c44d29742ac23638089f05" Dec 08 15:11:53 crc kubenswrapper[4894]: I1208 15:11:53.255850 4894 scope.go:117] "RemoveContainer" containerID="e3452f0a9e7d0db4ef898885ffadad684b50e329819d3cb90dd9d5f00bcf4f81" Dec 08 15:11:53 crc kubenswrapper[4894]: I1208 15:11:53.281673 4894 scope.go:117] "RemoveContainer" containerID="274b7c661c675a921bc3a18d3008a9578a6b50b96b3989dc752e5e691a8afc41" Dec 08 15:12:07 crc kubenswrapper[4894]: I1208 15:12:07.296310 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:12:07 crc kubenswrapper[4894]: I1208 15:12:07.296834 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:12:07 crc kubenswrapper[4894]: I1208 15:12:07.296889 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:12:07 crc kubenswrapper[4894]: I1208 15:12:07.299610 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:12:07 crc kubenswrapper[4894]: I1208 15:12:07.299724 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" gracePeriod=600 Dec 08 15:12:07 crc kubenswrapper[4894]: E1208 15:12:07.419029 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:12:08 crc kubenswrapper[4894]: I1208 15:12:08.310351 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" exitCode=0 Dec 08 15:12:08 crc kubenswrapper[4894]: I1208 15:12:08.310418 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa"} Dec 08 15:12:08 crc kubenswrapper[4894]: I1208 15:12:08.310723 4894 scope.go:117] "RemoveContainer" containerID="d9103dd6f195ccf766c15c1a9fa81340683956ac05b3d32c6e73e01c28e7fa11" Dec 08 15:12:08 crc kubenswrapper[4894]: I1208 15:12:08.311533 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:12:08 crc kubenswrapper[4894]: E1208 15:12:08.311879 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:12:17 crc kubenswrapper[4894]: I1208 15:12:17.394669 4894 generic.go:334] "Generic (PLEG): container finished" podID="86e41fca-6958-4b8a-84c0-851b75f8efea" containerID="8f41302a2a3dad69a1baf29bd8dec83d3563c78bace70935536ab3e9888a5c3f" exitCode=0 Dec 08 15:12:17 crc kubenswrapper[4894]: I1208 15:12:17.394758 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" event={"ID":"86e41fca-6958-4b8a-84c0-851b75f8efea","Type":"ContainerDied","Data":"8f41302a2a3dad69a1baf29bd8dec83d3563c78bace70935536ab3e9888a5c3f"} Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.936343 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.947646 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxfc2\" (UniqueName: \"kubernetes.io/projected/86e41fca-6958-4b8a-84c0-851b75f8efea-kube-api-access-kxfc2\") pod \"86e41fca-6958-4b8a-84c0-851b75f8efea\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.947811 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-ssh-key\") pod \"86e41fca-6958-4b8a-84c0-851b75f8efea\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.947905 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-bootstrap-combined-ca-bundle\") pod \"86e41fca-6958-4b8a-84c0-851b75f8efea\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.947956 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-inventory\") pod \"86e41fca-6958-4b8a-84c0-851b75f8efea\" (UID: \"86e41fca-6958-4b8a-84c0-851b75f8efea\") " Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.956172 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "86e41fca-6958-4b8a-84c0-851b75f8efea" (UID: "86e41fca-6958-4b8a-84c0-851b75f8efea"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.957641 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86e41fca-6958-4b8a-84c0-851b75f8efea-kube-api-access-kxfc2" (OuterVolumeSpecName: "kube-api-access-kxfc2") pod "86e41fca-6958-4b8a-84c0-851b75f8efea" (UID: "86e41fca-6958-4b8a-84c0-851b75f8efea"). InnerVolumeSpecName "kube-api-access-kxfc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.983168 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-inventory" (OuterVolumeSpecName: "inventory") pod "86e41fca-6958-4b8a-84c0-851b75f8efea" (UID: "86e41fca-6958-4b8a-84c0-851b75f8efea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:12:18 crc kubenswrapper[4894]: I1208 15:12:18.988018 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86e41fca-6958-4b8a-84c0-851b75f8efea" (UID: "86e41fca-6958-4b8a-84c0-851b75f8efea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.049896 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxfc2\" (UniqueName: \"kubernetes.io/projected/86e41fca-6958-4b8a-84c0-851b75f8efea-kube-api-access-kxfc2\") on node \"crc\" DevicePath \"\"" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.049928 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.049940 4894 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.049949 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86e41fca-6958-4b8a-84c0-851b75f8efea-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.209494 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:12:19 crc kubenswrapper[4894]: E1208 15:12:19.209808 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.413594 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" event={"ID":"86e41fca-6958-4b8a-84c0-851b75f8efea","Type":"ContainerDied","Data":"4bd80ab75e379aaa1e3824e913350b9f0bb5ea9ea011d4781f0ebb08c608bf27"} Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.413963 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bd80ab75e379aaa1e3824e913350b9f0bb5ea9ea011d4781f0ebb08c608bf27" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.413642 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.505363 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z"] Dec 08 15:12:19 crc kubenswrapper[4894]: E1208 15:12:19.505857 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86e41fca-6958-4b8a-84c0-851b75f8efea" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.505879 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="86e41fca-6958-4b8a-84c0-851b75f8efea" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 08 15:12:19 crc kubenswrapper[4894]: E1208 15:12:19.505890 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="extract-content" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.505897 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="extract-content" Dec 08 15:12:19 crc kubenswrapper[4894]: E1208 15:12:19.505920 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="extract-utilities" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.505927 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="extract-utilities" Dec 08 15:12:19 crc kubenswrapper[4894]: E1208 15:12:19.505942 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="registry-server" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.505948 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="registry-server" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.506112 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="86e41fca-6958-4b8a-84c0-851b75f8efea" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.506136 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a232a29c-ae10-425c-abda-7aff71396bcc" containerName="registry-server" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.506758 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.509987 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.510137 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.510238 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.510725 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.515530 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z"] Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.556733 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.556785 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.556888 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvz2v\" (UniqueName: \"kubernetes.io/projected/d2537804-7d9e-41e2-9224-58916974c998-kube-api-access-wvz2v\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.658544 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.658659 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.658860 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvz2v\" (UniqueName: \"kubernetes.io/projected/d2537804-7d9e-41e2-9224-58916974c998-kube-api-access-wvz2v\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.662574 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.663227 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.677238 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvz2v\" (UniqueName: \"kubernetes.io/projected/d2537804-7d9e-41e2-9224-58916974c998-kube-api-access-wvz2v\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:19 crc kubenswrapper[4894]: I1208 15:12:19.836973 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:12:20 crc kubenswrapper[4894]: I1208 15:12:20.414850 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z"] Dec 08 15:12:20 crc kubenswrapper[4894]: I1208 15:12:20.422138 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:12:21 crc kubenswrapper[4894]: I1208 15:12:21.437180 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" event={"ID":"d2537804-7d9e-41e2-9224-58916974c998","Type":"ContainerStarted","Data":"32b586f95798b59c6087ea17bed5ccdf31bb445bd3d91ffb3801fef3a48c9132"} Dec 08 15:12:21 crc kubenswrapper[4894]: I1208 15:12:21.437763 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" event={"ID":"d2537804-7d9e-41e2-9224-58916974c998","Type":"ContainerStarted","Data":"bde16bfdf3fe4eb9e93700b5ea6d268a7a6c8d72af98662ccc84ff6c20ab5bf9"} Dec 08 15:12:21 crc kubenswrapper[4894]: I1208 15:12:21.470254 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" podStartSLOduration=2.053832484 podStartE2EDuration="2.47023555s" podCreationTimestamp="2025-12-08 15:12:19 +0000 UTC" firstStartedPulling="2025-12-08 15:12:20.421917088 +0000 UTC m=+1561.521923193" lastFinishedPulling="2025-12-08 15:12:20.838320144 +0000 UTC m=+1561.938326259" observedRunningTime="2025-12-08 15:12:21.453427536 +0000 UTC m=+1562.553433651" watchObservedRunningTime="2025-12-08 15:12:21.47023555 +0000 UTC m=+1562.570241665" Dec 08 15:12:33 crc kubenswrapper[4894]: I1208 15:12:33.196900 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:12:33 crc kubenswrapper[4894]: E1208 15:12:33.197726 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.262999 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5xhjz"] Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.265409 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.284896 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cbdeb86-106a-426b-a285-ddfe70e7843a-catalog-content\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.284991 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dwgx\" (UniqueName: \"kubernetes.io/projected/4cbdeb86-106a-426b-a285-ddfe70e7843a-kube-api-access-5dwgx\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.286049 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cbdeb86-106a-426b-a285-ddfe70e7843a-utilities\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.290624 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5xhjz"] Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.388635 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cbdeb86-106a-426b-a285-ddfe70e7843a-catalog-content\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.389005 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dwgx\" (UniqueName: \"kubernetes.io/projected/4cbdeb86-106a-426b-a285-ddfe70e7843a-kube-api-access-5dwgx\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.389200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cbdeb86-106a-426b-a285-ddfe70e7843a-utilities\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.389479 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cbdeb86-106a-426b-a285-ddfe70e7843a-utilities\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.389600 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cbdeb86-106a-426b-a285-ddfe70e7843a-catalog-content\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.409662 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dwgx\" (UniqueName: \"kubernetes.io/projected/4cbdeb86-106a-426b-a285-ddfe70e7843a-kube-api-access-5dwgx\") pod \"redhat-operators-5xhjz\" (UID: \"4cbdeb86-106a-426b-a285-ddfe70e7843a\") " pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:45 crc kubenswrapper[4894]: I1208 15:12:45.588449 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:12:46 crc kubenswrapper[4894]: I1208 15:12:46.090336 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5xhjz"] Dec 08 15:12:46 crc kubenswrapper[4894]: I1208 15:12:46.196451 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:12:46 crc kubenswrapper[4894]: E1208 15:12:46.196752 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:12:46 crc kubenswrapper[4894]: I1208 15:12:46.691756 4894 generic.go:334] "Generic (PLEG): container finished" podID="4cbdeb86-106a-426b-a285-ddfe70e7843a" containerID="d822c0871247f66eae4bbe0e55ab274352c655a637727d268fc28d78bb44992c" exitCode=0 Dec 08 15:12:46 crc kubenswrapper[4894]: I1208 15:12:46.691899 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xhjz" event={"ID":"4cbdeb86-106a-426b-a285-ddfe70e7843a","Type":"ContainerDied","Data":"d822c0871247f66eae4bbe0e55ab274352c655a637727d268fc28d78bb44992c"} Dec 08 15:12:46 crc kubenswrapper[4894]: I1208 15:12:46.692359 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xhjz" event={"ID":"4cbdeb86-106a-426b-a285-ddfe70e7843a","Type":"ContainerStarted","Data":"fe6327fee44f6bda068202f4bf0d1e78c10ed2e497ea37b831aeac58b9d27efb"} Dec 08 15:12:53 crc kubenswrapper[4894]: I1208 15:12:53.400482 4894 scope.go:117] "RemoveContainer" containerID="a5de30d8ee82eda31a4dfcf3131e61c9e9f7cda63f16122fc505fa1a95dea11a" Dec 08 15:12:55 crc kubenswrapper[4894]: I1208 15:12:55.838071 4894 scope.go:117] "RemoveContainer" containerID="e9d6385bac32c049c7102f7d64837cc76fddc64ba7f1d26582fcd76ef5585017" Dec 08 15:12:56 crc kubenswrapper[4894]: I1208 15:12:56.790085 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xhjz" event={"ID":"4cbdeb86-106a-426b-a285-ddfe70e7843a","Type":"ContainerStarted","Data":"681cace3c31d390abfbba2de0d8fc58a48acbc17dc2cb5623aee773674754ac4"} Dec 08 15:12:58 crc kubenswrapper[4894]: I1208 15:12:58.197761 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:12:58 crc kubenswrapper[4894]: E1208 15:12:58.198052 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:12:58 crc kubenswrapper[4894]: I1208 15:12:58.813743 4894 generic.go:334] "Generic (PLEG): container finished" podID="4cbdeb86-106a-426b-a285-ddfe70e7843a" containerID="681cace3c31d390abfbba2de0d8fc58a48acbc17dc2cb5623aee773674754ac4" exitCode=0 Dec 08 15:12:58 crc kubenswrapper[4894]: I1208 15:12:58.813855 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xhjz" event={"ID":"4cbdeb86-106a-426b-a285-ddfe70e7843a","Type":"ContainerDied","Data":"681cace3c31d390abfbba2de0d8fc58a48acbc17dc2cb5623aee773674754ac4"} Dec 08 15:12:59 crc kubenswrapper[4894]: I1208 15:12:59.825218 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xhjz" event={"ID":"4cbdeb86-106a-426b-a285-ddfe70e7843a","Type":"ContainerStarted","Data":"23e141e422c474cef50ac04d5f0a1674ba339290d09a9d1dc445e6123aeca57d"} Dec 08 15:12:59 crc kubenswrapper[4894]: I1208 15:12:59.843129 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5xhjz" podStartSLOduration=2.03545563 podStartE2EDuration="14.84310694s" podCreationTimestamp="2025-12-08 15:12:45 +0000 UTC" firstStartedPulling="2025-12-08 15:12:46.69457027 +0000 UTC m=+1587.794576385" lastFinishedPulling="2025-12-08 15:12:59.50222158 +0000 UTC m=+1600.602227695" observedRunningTime="2025-12-08 15:12:59.841378357 +0000 UTC m=+1600.941384482" watchObservedRunningTime="2025-12-08 15:12:59.84310694 +0000 UTC m=+1600.943113055" Dec 08 15:13:05 crc kubenswrapper[4894]: I1208 15:13:05.589846 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:13:05 crc kubenswrapper[4894]: I1208 15:13:05.590542 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:13:05 crc kubenswrapper[4894]: I1208 15:13:05.655214 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:13:05 crc kubenswrapper[4894]: I1208 15:13:05.923520 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5xhjz" Dec 08 15:13:05 crc kubenswrapper[4894]: I1208 15:13:05.988973 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5xhjz"] Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.028297 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52vvt"] Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.028599 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-52vvt" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="registry-server" containerID="cri-o://0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d" gracePeriod=2 Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.481861 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.616145 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vssps\" (UniqueName: \"kubernetes.io/projected/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-kube-api-access-vssps\") pod \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.616311 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-utilities\") pod \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.616418 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-catalog-content\") pod \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\" (UID: \"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3\") " Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.616926 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-utilities" (OuterVolumeSpecName: "utilities") pod "7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" (UID: "7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.623671 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-kube-api-access-vssps" (OuterVolumeSpecName: "kube-api-access-vssps") pod "7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" (UID: "7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3"). InnerVolumeSpecName "kube-api-access-vssps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.718705 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.718745 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vssps\" (UniqueName: \"kubernetes.io/projected/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-kube-api-access-vssps\") on node \"crc\" DevicePath \"\"" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.754143 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" (UID: "7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.820377 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.886910 4894 generic.go:334] "Generic (PLEG): container finished" podID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerID="0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d" exitCode=0 Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.887296 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-52vvt" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.887348 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vvt" event={"ID":"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3","Type":"ContainerDied","Data":"0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d"} Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.887386 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-52vvt" event={"ID":"7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3","Type":"ContainerDied","Data":"3253988fdffbecf04ade87ec75761a2668d3582a6fa839d85745dbe95e6185bc"} Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.887404 4894 scope.go:117] "RemoveContainer" containerID="0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.913585 4894 scope.go:117] "RemoveContainer" containerID="f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.923567 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-52vvt"] Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.934427 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-52vvt"] Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.956182 4894 scope.go:117] "RemoveContainer" containerID="0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.990611 4894 scope.go:117] "RemoveContainer" containerID="0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d" Dec 08 15:13:06 crc kubenswrapper[4894]: E1208 15:13:06.991655 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d\": container with ID starting with 0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d not found: ID does not exist" containerID="0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.991759 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d"} err="failed to get container status \"0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d\": rpc error: code = NotFound desc = could not find container \"0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d\": container with ID starting with 0e6144c3f9ff5fe83854e7990c4ea0b187bda851b1e9004fdc7ed604fba8b11d not found: ID does not exist" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.991862 4894 scope.go:117] "RemoveContainer" containerID="f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474" Dec 08 15:13:06 crc kubenswrapper[4894]: E1208 15:13:06.992660 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474\": container with ID starting with f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474 not found: ID does not exist" containerID="f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.992742 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474"} err="failed to get container status \"f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474\": rpc error: code = NotFound desc = could not find container \"f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474\": container with ID starting with f56dcf2e8fb3f092ff3fac77a6adf1f2d3bf0d958156cce81077d66b5a50a474 not found: ID does not exist" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.992808 4894 scope.go:117] "RemoveContainer" containerID="0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9" Dec 08 15:13:06 crc kubenswrapper[4894]: E1208 15:13:06.994031 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9\": container with ID starting with 0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9 not found: ID does not exist" containerID="0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9" Dec 08 15:13:06 crc kubenswrapper[4894]: I1208 15:13:06.994074 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9"} err="failed to get container status \"0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9\": rpc error: code = NotFound desc = could not find container \"0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9\": container with ID starting with 0c7c6028a94965872c6c772ef6ed00c74a7dc5e63090f3e301b3e9148a5802a9 not found: ID does not exist" Dec 08 15:13:07 crc kubenswrapper[4894]: I1208 15:13:07.208379 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" path="/var/lib/kubelet/pods/7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3/volumes" Dec 08 15:13:13 crc kubenswrapper[4894]: I1208 15:13:13.196758 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:13:13 crc kubenswrapper[4894]: E1208 15:13:13.197578 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.058789 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-0cde-account-create-update-gbrxw"] Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.071512 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bgv76"] Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.082051 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b189-account-create-update-cg7wl"] Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.089299 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-87dgj"] Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.096708 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bgv76"] Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.103355 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b189-account-create-update-cg7wl"] Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.110291 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-87dgj"] Dec 08 15:13:14 crc kubenswrapper[4894]: I1208 15:13:14.117419 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-0cde-account-create-update-gbrxw"] Dec 08 15:13:15 crc kubenswrapper[4894]: I1208 15:13:15.219418 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28adf25f-022e-4386-b11e-1cd9fec44b3e" path="/var/lib/kubelet/pods/28adf25f-022e-4386-b11e-1cd9fec44b3e/volumes" Dec 08 15:13:15 crc kubenswrapper[4894]: I1208 15:13:15.220518 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56c30880-9aba-4ace-8bae-51c96dadd06e" path="/var/lib/kubelet/pods/56c30880-9aba-4ace-8bae-51c96dadd06e/volumes" Dec 08 15:13:15 crc kubenswrapper[4894]: I1208 15:13:15.221453 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acc10b3a-43ab-4a08-a981-d6dc24b9b5af" path="/var/lib/kubelet/pods/acc10b3a-43ab-4a08-a981-d6dc24b9b5af/volumes" Dec 08 15:13:15 crc kubenswrapper[4894]: I1208 15:13:15.222454 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae2689b5-0848-4b5f-8683-ecacf3f46dc8" path="/var/lib/kubelet/pods/ae2689b5-0848-4b5f-8683-ecacf3f46dc8/volumes" Dec 08 15:13:19 crc kubenswrapper[4894]: I1208 15:13:19.067840 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-2tlzl"] Dec 08 15:13:19 crc kubenswrapper[4894]: I1208 15:13:19.077125 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-79ea-account-create-update-66lff"] Dec 08 15:13:19 crc kubenswrapper[4894]: I1208 15:13:19.086312 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-2tlzl"] Dec 08 15:13:19 crc kubenswrapper[4894]: I1208 15:13:19.095462 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-79ea-account-create-update-66lff"] Dec 08 15:13:19 crc kubenswrapper[4894]: I1208 15:13:19.220115 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="739b42e0-365d-4eff-8724-d0de80c7d29f" path="/var/lib/kubelet/pods/739b42e0-365d-4eff-8724-d0de80c7d29f/volumes" Dec 08 15:13:19 crc kubenswrapper[4894]: I1208 15:13:19.223075 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb83cfcb-cfb9-4c84-919a-f4408f4962c5" path="/var/lib/kubelet/pods/eb83cfcb-cfb9-4c84-919a-f4408f4962c5/volumes" Dec 08 15:13:27 crc kubenswrapper[4894]: I1208 15:13:27.196566 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:13:27 crc kubenswrapper[4894]: E1208 15:13:27.197420 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:13:42 crc kubenswrapper[4894]: I1208 15:13:42.196436 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:13:42 crc kubenswrapper[4894]: E1208 15:13:42.197206 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:13:50 crc kubenswrapper[4894]: I1208 15:13:50.284530 4894 generic.go:334] "Generic (PLEG): container finished" podID="d2537804-7d9e-41e2-9224-58916974c998" containerID="32b586f95798b59c6087ea17bed5ccdf31bb445bd3d91ffb3801fef3a48c9132" exitCode=0 Dec 08 15:13:50 crc kubenswrapper[4894]: I1208 15:13:50.284625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" event={"ID":"d2537804-7d9e-41e2-9224-58916974c998","Type":"ContainerDied","Data":"32b586f95798b59c6087ea17bed5ccdf31bb445bd3d91ffb3801fef3a48c9132"} Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.043037 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-mld62"] Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.053015 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-mld62"] Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.207103 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd7c801c-c3b9-47cd-83ce-6a2de4083655" path="/var/lib/kubelet/pods/fd7c801c-c3b9-47cd-83ce-6a2de4083655/volumes" Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.673066 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.790925 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvz2v\" (UniqueName: \"kubernetes.io/projected/d2537804-7d9e-41e2-9224-58916974c998-kube-api-access-wvz2v\") pod \"d2537804-7d9e-41e2-9224-58916974c998\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.791040 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-inventory\") pod \"d2537804-7d9e-41e2-9224-58916974c998\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.791063 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-ssh-key\") pod \"d2537804-7d9e-41e2-9224-58916974c998\" (UID: \"d2537804-7d9e-41e2-9224-58916974c998\") " Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.796201 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2537804-7d9e-41e2-9224-58916974c998-kube-api-access-wvz2v" (OuterVolumeSpecName: "kube-api-access-wvz2v") pod "d2537804-7d9e-41e2-9224-58916974c998" (UID: "d2537804-7d9e-41e2-9224-58916974c998"). InnerVolumeSpecName "kube-api-access-wvz2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.817786 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-inventory" (OuterVolumeSpecName: "inventory") pod "d2537804-7d9e-41e2-9224-58916974c998" (UID: "d2537804-7d9e-41e2-9224-58916974c998"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.821042 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2537804-7d9e-41e2-9224-58916974c998" (UID: "d2537804-7d9e-41e2-9224-58916974c998"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.893774 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvz2v\" (UniqueName: \"kubernetes.io/projected/d2537804-7d9e-41e2-9224-58916974c998-kube-api-access-wvz2v\") on node \"crc\" DevicePath \"\"" Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.893806 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:13:51 crc kubenswrapper[4894]: I1208 15:13:51.893829 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2537804-7d9e-41e2-9224-58916974c998-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.028426 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d5f6-account-create-update-chx9n"] Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.039808 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d5f6-account-create-update-chx9n"] Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.303577 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" event={"ID":"d2537804-7d9e-41e2-9224-58916974c998","Type":"ContainerDied","Data":"bde16bfdf3fe4eb9e93700b5ea6d268a7a6c8d72af98662ccc84ff6c20ab5bf9"} Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.303894 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bde16bfdf3fe4eb9e93700b5ea6d268a7a6c8d72af98662ccc84ff6c20ab5bf9" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.303647 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.391297 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n"] Dec 08 15:13:52 crc kubenswrapper[4894]: E1208 15:13:52.391779 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="registry-server" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.391797 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="registry-server" Dec 08 15:13:52 crc kubenswrapper[4894]: E1208 15:13:52.391827 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="extract-content" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.391836 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="extract-content" Dec 08 15:13:52 crc kubenswrapper[4894]: E1208 15:13:52.391865 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2537804-7d9e-41e2-9224-58916974c998" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.391875 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2537804-7d9e-41e2-9224-58916974c998" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 08 15:13:52 crc kubenswrapper[4894]: E1208 15:13:52.391899 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="extract-utilities" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.391907 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="extract-utilities" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.392164 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2537804-7d9e-41e2-9224-58916974c998" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.392187 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a8f6f3b-c9a3-42c3-a006-8d3c0935aaa3" containerName="registry-server" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.393930 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.404071 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n"] Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.405041 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.405272 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.405478 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.406341 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.504395 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4mt9\" (UniqueName: \"kubernetes.io/projected/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-kube-api-access-s4mt9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.504514 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.504555 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.606086 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.606784 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.607155 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4mt9\" (UniqueName: \"kubernetes.io/projected/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-kube-api-access-s4mt9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.611151 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.611170 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.625206 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4mt9\" (UniqueName: \"kubernetes.io/projected/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-kube-api-access-s4mt9\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-x999n\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:52 crc kubenswrapper[4894]: I1208 15:13:52.733292 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:13:53 crc kubenswrapper[4894]: I1208 15:13:53.210407 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab8e2753-f67e-42fa-9374-1a51513d4e62" path="/var/lib/kubelet/pods/ab8e2753-f67e-42fa-9374-1a51513d4e62/volumes" Dec 08 15:13:53 crc kubenswrapper[4894]: I1208 15:13:53.280623 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n"] Dec 08 15:13:53 crc kubenswrapper[4894]: W1208 15:13:53.284519 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6f24a78_43c1_491e_98d0_3bd4d1d045eb.slice/crio-8923745f465583b927d547e57a1e77fe4e25502c0f1fdd0877e5cd87979a248a WatchSource:0}: Error finding container 8923745f465583b927d547e57a1e77fe4e25502c0f1fdd0877e5cd87979a248a: Status 404 returned error can't find the container with id 8923745f465583b927d547e57a1e77fe4e25502c0f1fdd0877e5cd87979a248a Dec 08 15:13:53 crc kubenswrapper[4894]: I1208 15:13:53.313677 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" event={"ID":"e6f24a78-43c1-491e-98d0-3bd4d1d045eb","Type":"ContainerStarted","Data":"8923745f465583b927d547e57a1e77fe4e25502c0f1fdd0877e5cd87979a248a"} Dec 08 15:13:54 crc kubenswrapper[4894]: I1208 15:13:54.322259 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" event={"ID":"e6f24a78-43c1-491e-98d0-3bd4d1d045eb","Type":"ContainerStarted","Data":"f7d4115c919b99ba026c00a7dc10e5c470d4fa7e122a02a484c279f704e77154"} Dec 08 15:13:54 crc kubenswrapper[4894]: I1208 15:13:54.347141 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" podStartSLOduration=1.848406427 podStartE2EDuration="2.347115855s" podCreationTimestamp="2025-12-08 15:13:52 +0000 UTC" firstStartedPulling="2025-12-08 15:13:53.28674931 +0000 UTC m=+1654.386755425" lastFinishedPulling="2025-12-08 15:13:53.785458738 +0000 UTC m=+1654.885464853" observedRunningTime="2025-12-08 15:13:54.334993435 +0000 UTC m=+1655.434999580" watchObservedRunningTime="2025-12-08 15:13:54.347115855 +0000 UTC m=+1655.447121990" Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.029647 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-tdflp"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.043879 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-tdflp"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.055663 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c7e9-account-create-update-qrjvs"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.063709 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c7e9-account-create-update-qrjvs"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.072134 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6e42-account-create-update-pbwqs"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.079867 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-stx47"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.087057 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6e42-account-create-update-pbwqs"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.094746 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-stx47"] Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.208960 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58032ce3-99e4-4cce-a421-7d427cc03cce" path="/var/lib/kubelet/pods/58032ce3-99e4-4cce-a421-7d427cc03cce/volumes" Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.209732 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65b6af4f-39e9-4afe-be5c-eaa60088ecde" path="/var/lib/kubelet/pods/65b6af4f-39e9-4afe-be5c-eaa60088ecde/volumes" Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.210421 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65" path="/var/lib/kubelet/pods/e76ce3d6-c2a1-40a1-93fc-bc4e4b1cef65/volumes" Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.211051 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10d506c-37b1-4046-99df-d106a244fed2" path="/var/lib/kubelet/pods/f10d506c-37b1-4046-99df-d106a244fed2/volumes" Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.924693 4894 scope.go:117] "RemoveContainer" containerID="077ecace2cc9e820c860d2d8db70dda18705dcf42fde3c7e0d4239a26a2741de" Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.947058 4894 scope.go:117] "RemoveContainer" containerID="ec96aecf3474d70067e2913cacca6f962cf9daf3dced6deb5dc000469be0e303" Dec 08 15:13:55 crc kubenswrapper[4894]: I1208 15:13:55.992949 4894 scope.go:117] "RemoveContainer" containerID="e69f3993e82eea147e8983c144e2d29b670963d1602072fb5e21cfd677a72259" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.033533 4894 scope.go:117] "RemoveContainer" containerID="abd82b1ce872f8717d6a0f43c585d8bfdca536bac15e496575336c6637518624" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.084268 4894 scope.go:117] "RemoveContainer" containerID="1448200b24854f7d9e77c123734c59d21f22129bd43dca58f3b808e4d1e980fd" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.131338 4894 scope.go:117] "RemoveContainer" containerID="f12a4b2aa1ee4b7c4f3373e8167c434c3bb616552e964f4b9e0d722b5a57b9a4" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.174313 4894 scope.go:117] "RemoveContainer" containerID="6335e9953b6c8e00d68c28474beaa8b2235a73cfce70533d4408b8c4ac92faf2" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.193725 4894 scope.go:117] "RemoveContainer" containerID="25d99ea25092e78c89ee8bbdc8bf761f07ba80d32f6477c77a453ac7ec0385b9" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.196997 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:13:56 crc kubenswrapper[4894]: E1208 15:13:56.197305 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.214555 4894 scope.go:117] "RemoveContainer" containerID="683b976da4ad77cc67f2d9d7983262120b45b1f63e3fcea35055d68dfd6653ef" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.232413 4894 scope.go:117] "RemoveContainer" containerID="02d48033cc2628c0375de65f65e42a41c4853aab4435f51634d3328daffb06ff" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.258499 4894 scope.go:117] "RemoveContainer" containerID="8ca97fea586b6476c6019ed0179f40675e623cd73290e180b6707f6d74f23592" Dec 08 15:13:56 crc kubenswrapper[4894]: I1208 15:13:56.276789 4894 scope.go:117] "RemoveContainer" containerID="ac86e1b5e294420010e0f40667ad6ff07cf90eb34032ab45e3720e31a8e6aa79" Dec 08 15:14:00 crc kubenswrapper[4894]: I1208 15:14:00.042753 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ztqbr"] Dec 08 15:14:00 crc kubenswrapper[4894]: I1208 15:14:00.053521 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ztqbr"] Dec 08 15:14:01 crc kubenswrapper[4894]: I1208 15:14:01.212760 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="473b20c5-eb88-4b48-b198-f828dca0e7c0" path="/var/lib/kubelet/pods/473b20c5-eb88-4b48-b198-f828dca0e7c0/volumes" Dec 08 15:14:11 crc kubenswrapper[4894]: I1208 15:14:11.196644 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:14:11 crc kubenswrapper[4894]: E1208 15:14:11.197318 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:14:23 crc kubenswrapper[4894]: I1208 15:14:23.196796 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:14:23 crc kubenswrapper[4894]: E1208 15:14:23.197563 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:14:32 crc kubenswrapper[4894]: I1208 15:14:32.043877 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-htqq9"] Dec 08 15:14:32 crc kubenswrapper[4894]: I1208 15:14:32.052413 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-htqq9"] Dec 08 15:14:33 crc kubenswrapper[4894]: I1208 15:14:33.211413 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5041f7-6cf0-46c8-8fc4-9616005d46b7" path="/var/lib/kubelet/pods/bc5041f7-6cf0-46c8-8fc4-9616005d46b7/volumes" Dec 08 15:14:34 crc kubenswrapper[4894]: I1208 15:14:34.197421 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:14:34 crc kubenswrapper[4894]: E1208 15:14:34.197886 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:14:35 crc kubenswrapper[4894]: I1208 15:14:35.043572 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kzgzh"] Dec 08 15:14:35 crc kubenswrapper[4894]: I1208 15:14:35.051533 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kzgzh"] Dec 08 15:14:35 crc kubenswrapper[4894]: I1208 15:14:35.215880 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f" path="/var/lib/kubelet/pods/bfdaa6a7-4991-4cc0-b19c-1d2e0f067f4f/volumes" Dec 08 15:14:37 crc kubenswrapper[4894]: I1208 15:14:37.045693 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-g4klr"] Dec 08 15:14:37 crc kubenswrapper[4894]: I1208 15:14:37.058325 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-g4klr"] Dec 08 15:14:37 crc kubenswrapper[4894]: I1208 15:14:37.208664 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="195700ca-d25a-470a-a6c4-73d4f4133b95" path="/var/lib/kubelet/pods/195700ca-d25a-470a-a6c4-73d4f4133b95/volumes" Dec 08 15:14:40 crc kubenswrapper[4894]: I1208 15:14:40.038725 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-c642j"] Dec 08 15:14:40 crc kubenswrapper[4894]: I1208 15:14:40.047561 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-c642j"] Dec 08 15:14:41 crc kubenswrapper[4894]: I1208 15:14:41.210340 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a9427b-9569-447d-aa60-e3be638e2d4c" path="/var/lib/kubelet/pods/77a9427b-9569-447d-aa60-e3be638e2d4c/volumes" Dec 08 15:14:46 crc kubenswrapper[4894]: I1208 15:14:46.197024 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:14:46 crc kubenswrapper[4894]: E1208 15:14:46.197797 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.040618 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-7b8gb"] Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.050026 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7nxjg"] Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.059234 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-7b8gb"] Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.069220 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7nxjg"] Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.488584 4894 scope.go:117] "RemoveContainer" containerID="d9a002ecf2a43b769b650f5eaf45af28c9c5bcf093c16c5beffca14bd9a79304" Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.525902 4894 scope.go:117] "RemoveContainer" containerID="e366fbf43e4ab9f33b6bccddd40b102659722fc10c6fd28229bb934ee8d072e3" Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.577514 4894 scope.go:117] "RemoveContainer" containerID="491d9ef680de2ba394919a36913aee194bc87821c18ce36cd7bff21bf8cec0f5" Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.613110 4894 scope.go:117] "RemoveContainer" containerID="faa9d20bccbea00e7a42702e051752cab4d1e85396f3a5ed06a8ef428d55bb04" Dec 08 15:14:56 crc kubenswrapper[4894]: I1208 15:14:56.663649 4894 scope.go:117] "RemoveContainer" containerID="90aebf031ec3f48a436b2122f40dd850b0eeee78920dc26741dd010da6c2f482" Dec 08 15:14:57 crc kubenswrapper[4894]: I1208 15:14:57.212263 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06177501-5baa-49e0-94bd-157c6f079c05" path="/var/lib/kubelet/pods/06177501-5baa-49e0-94bd-157c6f079c05/volumes" Dec 08 15:14:57 crc kubenswrapper[4894]: I1208 15:14:57.214757 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85f57102-ceaf-484a-a0b3-c1ecf324339d" path="/var/lib/kubelet/pods/85f57102-ceaf-484a-a0b3-c1ecf324339d/volumes" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.148947 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc"] Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.151001 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.155026 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.155066 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.159786 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc"] Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.198919 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:15:00 crc kubenswrapper[4894]: E1208 15:15:00.199162 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.221770 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86f4a083-9b0f-437d-bffb-dfb8037c9c82-secret-volume\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.221962 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86f4a083-9b0f-437d-bffb-dfb8037c9c82-config-volume\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.222311 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcg9d\" (UniqueName: \"kubernetes.io/projected/86f4a083-9b0f-437d-bffb-dfb8037c9c82-kube-api-access-jcg9d\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.324367 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86f4a083-9b0f-437d-bffb-dfb8037c9c82-secret-volume\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.324454 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86f4a083-9b0f-437d-bffb-dfb8037c9c82-config-volume\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.324577 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcg9d\" (UniqueName: \"kubernetes.io/projected/86f4a083-9b0f-437d-bffb-dfb8037c9c82-kube-api-access-jcg9d\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.326522 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86f4a083-9b0f-437d-bffb-dfb8037c9c82-config-volume\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.332232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86f4a083-9b0f-437d-bffb-dfb8037c9c82-secret-volume\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.340479 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcg9d\" (UniqueName: \"kubernetes.io/projected/86f4a083-9b0f-437d-bffb-dfb8037c9c82-kube-api-access-jcg9d\") pod \"collect-profiles-29420115-mqsrc\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.484309 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:00 crc kubenswrapper[4894]: I1208 15:15:00.917731 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc"] Dec 08 15:15:01 crc kubenswrapper[4894]: I1208 15:15:01.945628 4894 generic.go:334] "Generic (PLEG): container finished" podID="86f4a083-9b0f-437d-bffb-dfb8037c9c82" containerID="a4c1064b90e63479ecf32318992c8462c991dafb00f2cc353d6a64530f031c64" exitCode=0 Dec 08 15:15:01 crc kubenswrapper[4894]: I1208 15:15:01.946199 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" event={"ID":"86f4a083-9b0f-437d-bffb-dfb8037c9c82","Type":"ContainerDied","Data":"a4c1064b90e63479ecf32318992c8462c991dafb00f2cc353d6a64530f031c64"} Dec 08 15:15:01 crc kubenswrapper[4894]: I1208 15:15:01.946233 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" event={"ID":"86f4a083-9b0f-437d-bffb-dfb8037c9c82","Type":"ContainerStarted","Data":"9c604d7c658dae3033f9f5e745f014e94a96d94300bc4d9f597f9c8a24552213"} Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.346906 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.400780 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86f4a083-9b0f-437d-bffb-dfb8037c9c82-secret-volume\") pod \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.401017 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86f4a083-9b0f-437d-bffb-dfb8037c9c82-config-volume\") pod \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.402124 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86f4a083-9b0f-437d-bffb-dfb8037c9c82-config-volume" (OuterVolumeSpecName: "config-volume") pod "86f4a083-9b0f-437d-bffb-dfb8037c9c82" (UID: "86f4a083-9b0f-437d-bffb-dfb8037c9c82"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.401048 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcg9d\" (UniqueName: \"kubernetes.io/projected/86f4a083-9b0f-437d-bffb-dfb8037c9c82-kube-api-access-jcg9d\") pod \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\" (UID: \"86f4a083-9b0f-437d-bffb-dfb8037c9c82\") " Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.403001 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86f4a083-9b0f-437d-bffb-dfb8037c9c82-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.411052 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86f4a083-9b0f-437d-bffb-dfb8037c9c82-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "86f4a083-9b0f-437d-bffb-dfb8037c9c82" (UID: "86f4a083-9b0f-437d-bffb-dfb8037c9c82"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.411518 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f4a083-9b0f-437d-bffb-dfb8037c9c82-kube-api-access-jcg9d" (OuterVolumeSpecName: "kube-api-access-jcg9d") pod "86f4a083-9b0f-437d-bffb-dfb8037c9c82" (UID: "86f4a083-9b0f-437d-bffb-dfb8037c9c82"). InnerVolumeSpecName "kube-api-access-jcg9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.505469 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86f4a083-9b0f-437d-bffb-dfb8037c9c82-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.505507 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcg9d\" (UniqueName: \"kubernetes.io/projected/86f4a083-9b0f-437d-bffb-dfb8037c9c82-kube-api-access-jcg9d\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.966627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" event={"ID":"86f4a083-9b0f-437d-bffb-dfb8037c9c82","Type":"ContainerDied","Data":"9c604d7c658dae3033f9f5e745f014e94a96d94300bc4d9f597f9c8a24552213"} Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.967006 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c604d7c658dae3033f9f5e745f014e94a96d94300bc4d9f597f9c8a24552213" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.966648 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420115-mqsrc" Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.968966 4894 generic.go:334] "Generic (PLEG): container finished" podID="e6f24a78-43c1-491e-98d0-3bd4d1d045eb" containerID="f7d4115c919b99ba026c00a7dc10e5c470d4fa7e122a02a484c279f704e77154" exitCode=0 Dec 08 15:15:03 crc kubenswrapper[4894]: I1208 15:15:03.969001 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" event={"ID":"e6f24a78-43c1-491e-98d0-3bd4d1d045eb","Type":"ContainerDied","Data":"f7d4115c919b99ba026c00a7dc10e5c470d4fa7e122a02a484c279f704e77154"} Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.387319 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.442154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4mt9\" (UniqueName: \"kubernetes.io/projected/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-kube-api-access-s4mt9\") pod \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.442427 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-inventory\") pod \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.442571 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-ssh-key\") pod \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\" (UID: \"e6f24a78-43c1-491e-98d0-3bd4d1d045eb\") " Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.447810 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-kube-api-access-s4mt9" (OuterVolumeSpecName: "kube-api-access-s4mt9") pod "e6f24a78-43c1-491e-98d0-3bd4d1d045eb" (UID: "e6f24a78-43c1-491e-98d0-3bd4d1d045eb"). InnerVolumeSpecName "kube-api-access-s4mt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.470367 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e6f24a78-43c1-491e-98d0-3bd4d1d045eb" (UID: "e6f24a78-43c1-491e-98d0-3bd4d1d045eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.479113 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-inventory" (OuterVolumeSpecName: "inventory") pod "e6f24a78-43c1-491e-98d0-3bd4d1d045eb" (UID: "e6f24a78-43c1-491e-98d0-3bd4d1d045eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.546501 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.546558 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4mt9\" (UniqueName: \"kubernetes.io/projected/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-kube-api-access-s4mt9\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.546583 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6f24a78-43c1-491e-98d0-3bd4d1d045eb-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.987125 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" event={"ID":"e6f24a78-43c1-491e-98d0-3bd4d1d045eb","Type":"ContainerDied","Data":"8923745f465583b927d547e57a1e77fe4e25502c0f1fdd0877e5cd87979a248a"} Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.987153 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-x999n" Dec 08 15:15:05 crc kubenswrapper[4894]: I1208 15:15:05.987159 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8923745f465583b927d547e57a1e77fe4e25502c0f1fdd0877e5cd87979a248a" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.062781 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj"] Dec 08 15:15:06 crc kubenswrapper[4894]: E1208 15:15:06.065483 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6f24a78-43c1-491e-98d0-3bd4d1d045eb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.065518 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6f24a78-43c1-491e-98d0-3bd4d1d045eb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:06 crc kubenswrapper[4894]: E1208 15:15:06.065545 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f4a083-9b0f-437d-bffb-dfb8037c9c82" containerName="collect-profiles" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.065552 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f4a083-9b0f-437d-bffb-dfb8037c9c82" containerName="collect-profiles" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.065792 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6f24a78-43c1-491e-98d0-3bd4d1d045eb" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.065833 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f4a083-9b0f-437d-bffb-dfb8037c9c82" containerName="collect-profiles" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.066637 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.071401 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.071766 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.071920 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.072074 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.078443 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj"] Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.154860 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.154930 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njlmq\" (UniqueName: \"kubernetes.io/projected/18936dba-63ba-44de-954c-c65df5b5743a-kube-api-access-njlmq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.155051 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.256412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.256560 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.256584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njlmq\" (UniqueName: \"kubernetes.io/projected/18936dba-63ba-44de-954c-c65df5b5743a-kube-api-access-njlmq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.261864 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.263658 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.280404 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njlmq\" (UniqueName: \"kubernetes.io/projected/18936dba-63ba-44de-954c-c65df5b5743a-kube-api-access-njlmq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.386598 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.921186 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj"] Dec 08 15:15:06 crc kubenswrapper[4894]: W1208 15:15:06.932350 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18936dba_63ba_44de_954c_c65df5b5743a.slice/crio-b57197febc2184e226c403b000f73f06844844bc6fc1c8a659aca9da5df2c67c WatchSource:0}: Error finding container b57197febc2184e226c403b000f73f06844844bc6fc1c8a659aca9da5df2c67c: Status 404 returned error can't find the container with id b57197febc2184e226c403b000f73f06844844bc6fc1c8a659aca9da5df2c67c Dec 08 15:15:06 crc kubenswrapper[4894]: I1208 15:15:06.997197 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" event={"ID":"18936dba-63ba-44de-954c-c65df5b5743a","Type":"ContainerStarted","Data":"b57197febc2184e226c403b000f73f06844844bc6fc1c8a659aca9da5df2c67c"} Dec 08 15:15:08 crc kubenswrapper[4894]: I1208 15:15:08.005476 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" event={"ID":"18936dba-63ba-44de-954c-c65df5b5743a","Type":"ContainerStarted","Data":"0cbc2801f85a7f1eeda4932e9fed57b5a632d4c9397163c17b7bd7cfb54acfc6"} Dec 08 15:15:08 crc kubenswrapper[4894]: I1208 15:15:08.032095 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" podStartSLOduration=1.514119041 podStartE2EDuration="2.032078661s" podCreationTimestamp="2025-12-08 15:15:06 +0000 UTC" firstStartedPulling="2025-12-08 15:15:06.935093152 +0000 UTC m=+1728.035099257" lastFinishedPulling="2025-12-08 15:15:07.453052762 +0000 UTC m=+1728.553058877" observedRunningTime="2025-12-08 15:15:08.025908329 +0000 UTC m=+1729.125914454" watchObservedRunningTime="2025-12-08 15:15:08.032078661 +0000 UTC m=+1729.132084776" Dec 08 15:15:13 crc kubenswrapper[4894]: I1208 15:15:13.057419 4894 generic.go:334] "Generic (PLEG): container finished" podID="18936dba-63ba-44de-954c-c65df5b5743a" containerID="0cbc2801f85a7f1eeda4932e9fed57b5a632d4c9397163c17b7bd7cfb54acfc6" exitCode=0 Dec 08 15:15:13 crc kubenswrapper[4894]: I1208 15:15:13.058033 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" event={"ID":"18936dba-63ba-44de-954c-c65df5b5743a","Type":"ContainerDied","Data":"0cbc2801f85a7f1eeda4932e9fed57b5a632d4c9397163c17b7bd7cfb54acfc6"} Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.465648 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.617020 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njlmq\" (UniqueName: \"kubernetes.io/projected/18936dba-63ba-44de-954c-c65df5b5743a-kube-api-access-njlmq\") pod \"18936dba-63ba-44de-954c-c65df5b5743a\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.617368 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-inventory\") pod \"18936dba-63ba-44de-954c-c65df5b5743a\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.617519 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-ssh-key\") pod \"18936dba-63ba-44de-954c-c65df5b5743a\" (UID: \"18936dba-63ba-44de-954c-c65df5b5743a\") " Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.626570 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18936dba-63ba-44de-954c-c65df5b5743a-kube-api-access-njlmq" (OuterVolumeSpecName: "kube-api-access-njlmq") pod "18936dba-63ba-44de-954c-c65df5b5743a" (UID: "18936dba-63ba-44de-954c-c65df5b5743a"). InnerVolumeSpecName "kube-api-access-njlmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.647967 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-inventory" (OuterVolumeSpecName: "inventory") pod "18936dba-63ba-44de-954c-c65df5b5743a" (UID: "18936dba-63ba-44de-954c-c65df5b5743a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.667008 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18936dba-63ba-44de-954c-c65df5b5743a" (UID: "18936dba-63ba-44de-954c-c65df5b5743a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.719546 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njlmq\" (UniqueName: \"kubernetes.io/projected/18936dba-63ba-44de-954c-c65df5b5743a-kube-api-access-njlmq\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.719797 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:14 crc kubenswrapper[4894]: I1208 15:15:14.720047 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18936dba-63ba-44de-954c-c65df5b5743a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.075565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" event={"ID":"18936dba-63ba-44de-954c-c65df5b5743a","Type":"ContainerDied","Data":"b57197febc2184e226c403b000f73f06844844bc6fc1c8a659aca9da5df2c67c"} Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.075618 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b57197febc2184e226c403b000f73f06844844bc6fc1c8a659aca9da5df2c67c" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.075756 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.163173 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf"] Dec 08 15:15:15 crc kubenswrapper[4894]: E1208 15:15:15.163637 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18936dba-63ba-44de-954c-c65df5b5743a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.163655 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="18936dba-63ba-44de-954c-c65df5b5743a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.163943 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="18936dba-63ba-44de-954c-c65df5b5743a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.164836 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.169381 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.169530 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.169694 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.173225 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf"] Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.173382 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.196701 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:15:15 crc kubenswrapper[4894]: E1208 15:15:15.197075 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.331724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8287\" (UniqueName: \"kubernetes.io/projected/dff601df-c402-489d-b878-a8aaa7a32909-kube-api-access-p8287\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.332257 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.332683 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.434356 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.434430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8287\" (UniqueName: \"kubernetes.io/projected/dff601df-c402-489d-b878-a8aaa7a32909-kube-api-access-p8287\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.434536 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.438527 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.439270 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.452474 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8287\" (UniqueName: \"kubernetes.io/projected/dff601df-c402-489d-b878-a8aaa7a32909-kube-api-access-p8287\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vcccf\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:15 crc kubenswrapper[4894]: I1208 15:15:15.485984 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:16 crc kubenswrapper[4894]: I1208 15:15:16.052078 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf"] Dec 08 15:15:16 crc kubenswrapper[4894]: I1208 15:15:16.084729 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" event={"ID":"dff601df-c402-489d-b878-a8aaa7a32909","Type":"ContainerStarted","Data":"b84d47477213387bbb99b1e4a9a2e81e3f0e3d6bd61e72b7ff7b123be5b72d0a"} Dec 08 15:15:17 crc kubenswrapper[4894]: I1208 15:15:17.095450 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" event={"ID":"dff601df-c402-489d-b878-a8aaa7a32909","Type":"ContainerStarted","Data":"f2942d4e6a538eb18f3c3cd5b55f94c975ac36a5c8a5273cb4bf716bff7ddfab"} Dec 08 15:15:17 crc kubenswrapper[4894]: I1208 15:15:17.115737 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" podStartSLOduration=1.664733156 podStartE2EDuration="2.115720482s" podCreationTimestamp="2025-12-08 15:15:15 +0000 UTC" firstStartedPulling="2025-12-08 15:15:16.052804098 +0000 UTC m=+1737.152810213" lastFinishedPulling="2025-12-08 15:15:16.503791424 +0000 UTC m=+1737.603797539" observedRunningTime="2025-12-08 15:15:17.10956557 +0000 UTC m=+1738.209571695" watchObservedRunningTime="2025-12-08 15:15:17.115720482 +0000 UTC m=+1738.215726597" Dec 08 15:15:29 crc kubenswrapper[4894]: I1208 15:15:29.202718 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:15:29 crc kubenswrapper[4894]: E1208 15:15:29.203525 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.754203 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jrdms"] Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.756659 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.776870 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jrdms"] Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.889165 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-catalog-content\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.889212 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-utilities\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.889806 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25l9p\" (UniqueName: \"kubernetes.io/projected/2efacc1c-1a99-4105-8034-b7554d1013f9-kube-api-access-25l9p\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.992240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25l9p\" (UniqueName: \"kubernetes.io/projected/2efacc1c-1a99-4105-8034-b7554d1013f9-kube-api-access-25l9p\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.992315 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-catalog-content\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.992338 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-utilities\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.992784 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-utilities\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:38 crc kubenswrapper[4894]: I1208 15:15:38.992876 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-catalog-content\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:39 crc kubenswrapper[4894]: I1208 15:15:39.010663 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25l9p\" (UniqueName: \"kubernetes.io/projected/2efacc1c-1a99-4105-8034-b7554d1013f9-kube-api-access-25l9p\") pod \"certified-operators-jrdms\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:39 crc kubenswrapper[4894]: I1208 15:15:39.086281 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:39 crc kubenswrapper[4894]: I1208 15:15:39.567317 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jrdms"] Dec 08 15:15:40 crc kubenswrapper[4894]: I1208 15:15:40.329589 4894 generic.go:334] "Generic (PLEG): container finished" podID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerID="5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764" exitCode=0 Dec 08 15:15:40 crc kubenswrapper[4894]: I1208 15:15:40.329928 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrdms" event={"ID":"2efacc1c-1a99-4105-8034-b7554d1013f9","Type":"ContainerDied","Data":"5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764"} Dec 08 15:15:40 crc kubenswrapper[4894]: I1208 15:15:40.329966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrdms" event={"ID":"2efacc1c-1a99-4105-8034-b7554d1013f9","Type":"ContainerStarted","Data":"16a410f21061061dcc56a4039724856e49fa475c811a155f774f7d9e973c3a8c"} Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.340563 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrdms" event={"ID":"2efacc1c-1a99-4105-8034-b7554d1013f9","Type":"ContainerStarted","Data":"c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0"} Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.742475 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nt99r"] Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.744705 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.756688 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4slc\" (UniqueName: \"kubernetes.io/projected/98c017a5-0df0-4e05-a7f8-a2d150910c44-kube-api-access-n4slc\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.756770 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-utilities\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.757047 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-catalog-content\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.761968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nt99r"] Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.858856 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-catalog-content\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.858969 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4slc\" (UniqueName: \"kubernetes.io/projected/98c017a5-0df0-4e05-a7f8-a2d150910c44-kube-api-access-n4slc\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.859001 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-utilities\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.859451 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-utilities\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.860022 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-catalog-content\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:41 crc kubenswrapper[4894]: I1208 15:15:41.881888 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4slc\" (UniqueName: \"kubernetes.io/projected/98c017a5-0df0-4e05-a7f8-a2d150910c44-kube-api-access-n4slc\") pod \"redhat-marketplace-nt99r\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:42 crc kubenswrapper[4894]: I1208 15:15:42.071326 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:42 crc kubenswrapper[4894]: I1208 15:15:42.351386 4894 generic.go:334] "Generic (PLEG): container finished" podID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerID="c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0" exitCode=0 Dec 08 15:15:42 crc kubenswrapper[4894]: I1208 15:15:42.351700 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrdms" event={"ID":"2efacc1c-1a99-4105-8034-b7554d1013f9","Type":"ContainerDied","Data":"c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0"} Dec 08 15:15:42 crc kubenswrapper[4894]: W1208 15:15:42.589041 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98c017a5_0df0_4e05_a7f8_a2d150910c44.slice/crio-64a49d539fedfd41c2b6ea10d010f9f62c3db82764ee76a9dada83954fff1345 WatchSource:0}: Error finding container 64a49d539fedfd41c2b6ea10d010f9f62c3db82764ee76a9dada83954fff1345: Status 404 returned error can't find the container with id 64a49d539fedfd41c2b6ea10d010f9f62c3db82764ee76a9dada83954fff1345 Dec 08 15:15:42 crc kubenswrapper[4894]: I1208 15:15:42.592499 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nt99r"] Dec 08 15:15:43 crc kubenswrapper[4894]: I1208 15:15:43.361476 4894 generic.go:334] "Generic (PLEG): container finished" podID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerID="ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1" exitCode=0 Dec 08 15:15:43 crc kubenswrapper[4894]: I1208 15:15:43.361681 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nt99r" event={"ID":"98c017a5-0df0-4e05-a7f8-a2d150910c44","Type":"ContainerDied","Data":"ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1"} Dec 08 15:15:43 crc kubenswrapper[4894]: I1208 15:15:43.361865 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nt99r" event={"ID":"98c017a5-0df0-4e05-a7f8-a2d150910c44","Type":"ContainerStarted","Data":"64a49d539fedfd41c2b6ea10d010f9f62c3db82764ee76a9dada83954fff1345"} Dec 08 15:15:43 crc kubenswrapper[4894]: I1208 15:15:43.364048 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrdms" event={"ID":"2efacc1c-1a99-4105-8034-b7554d1013f9","Type":"ContainerStarted","Data":"bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159"} Dec 08 15:15:43 crc kubenswrapper[4894]: I1208 15:15:43.414094 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jrdms" podStartSLOduration=2.91797094 podStartE2EDuration="5.414073442s" podCreationTimestamp="2025-12-08 15:15:38 +0000 UTC" firstStartedPulling="2025-12-08 15:15:40.332577012 +0000 UTC m=+1761.432583137" lastFinishedPulling="2025-12-08 15:15:42.828679524 +0000 UTC m=+1763.928685639" observedRunningTime="2025-12-08 15:15:43.400289191 +0000 UTC m=+1764.500295316" watchObservedRunningTime="2025-12-08 15:15:43.414073442 +0000 UTC m=+1764.514079567" Dec 08 15:15:44 crc kubenswrapper[4894]: I1208 15:15:44.041718 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-nf84z"] Dec 08 15:15:44 crc kubenswrapper[4894]: I1208 15:15:44.055266 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-nf84z"] Dec 08 15:15:44 crc kubenswrapper[4894]: I1208 15:15:44.196639 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:15:44 crc kubenswrapper[4894]: E1208 15:15:44.196913 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.026949 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-7hskq"] Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.034893 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c9d7-account-create-update-fkrpz"] Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.044149 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-7hskq"] Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.053154 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c9d7-account-create-update-fkrpz"] Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.067882 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8c84-account-create-update-vrjkn"] Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.080310 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8c84-account-create-update-vrjkn"] Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.208740 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0062ed70-9c1e-4201-85bd-9a01061bdfa0" path="/var/lib/kubelet/pods/0062ed70-9c1e-4201-85bd-9a01061bdfa0/volumes" Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.209520 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147a0472-ae65-499b-ad28-c5111f7ba6b2" path="/var/lib/kubelet/pods/147a0472-ae65-499b-ad28-c5111f7ba6b2/volumes" Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.210089 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="552bb30e-121a-4de7-ab1f-b9d4ae181482" path="/var/lib/kubelet/pods/552bb30e-121a-4de7-ab1f-b9d4ae181482/volumes" Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.210656 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cceeee5-f000-4e96-a019-ef0f1c1e9b09" path="/var/lib/kubelet/pods/8cceeee5-f000-4e96-a019-ef0f1c1e9b09/volumes" Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.384258 4894 generic.go:334] "Generic (PLEG): container finished" podID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerID="dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928" exitCode=0 Dec 08 15:15:45 crc kubenswrapper[4894]: I1208 15:15:45.384313 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nt99r" event={"ID":"98c017a5-0df0-4e05-a7f8-a2d150910c44","Type":"ContainerDied","Data":"dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928"} Dec 08 15:15:46 crc kubenswrapper[4894]: I1208 15:15:46.024248 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-2gjd8"] Dec 08 15:15:46 crc kubenswrapper[4894]: I1208 15:15:46.033235 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-2gjd8"] Dec 08 15:15:46 crc kubenswrapper[4894]: I1208 15:15:46.046156 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-024d-account-create-update-6zh4q"] Dec 08 15:15:46 crc kubenswrapper[4894]: I1208 15:15:46.054588 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-024d-account-create-update-6zh4q"] Dec 08 15:15:46 crc kubenswrapper[4894]: I1208 15:15:46.393913 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nt99r" event={"ID":"98c017a5-0df0-4e05-a7f8-a2d150910c44","Type":"ContainerStarted","Data":"35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8"} Dec 08 15:15:46 crc kubenswrapper[4894]: I1208 15:15:46.413323 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nt99r" podStartSLOduration=2.922611487 podStartE2EDuration="5.41330736s" podCreationTimestamp="2025-12-08 15:15:41 +0000 UTC" firstStartedPulling="2025-12-08 15:15:43.363397177 +0000 UTC m=+1764.463403292" lastFinishedPulling="2025-12-08 15:15:45.85409302 +0000 UTC m=+1766.954099165" observedRunningTime="2025-12-08 15:15:46.412160754 +0000 UTC m=+1767.512166869" watchObservedRunningTime="2025-12-08 15:15:46.41330736 +0000 UTC m=+1767.513313475" Dec 08 15:15:47 crc kubenswrapper[4894]: I1208 15:15:47.208150 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="374fd9cd-a481-46ba-80b5-6024953cd68d" path="/var/lib/kubelet/pods/374fd9cd-a481-46ba-80b5-6024953cd68d/volumes" Dec 08 15:15:47 crc kubenswrapper[4894]: I1208 15:15:47.209187 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a868f1eb-3d98-45c7-abaf-775e4f4e7db3" path="/var/lib/kubelet/pods/a868f1eb-3d98-45c7-abaf-775e4f4e7db3/volumes" Dec 08 15:15:49 crc kubenswrapper[4894]: I1208 15:15:49.086713 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:49 crc kubenswrapper[4894]: I1208 15:15:49.087449 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:49 crc kubenswrapper[4894]: I1208 15:15:49.191376 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:49 crc kubenswrapper[4894]: I1208 15:15:49.464498 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:50 crc kubenswrapper[4894]: I1208 15:15:50.737981 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jrdms"] Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.072234 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.072557 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.117499 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.442622 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jrdms" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="registry-server" containerID="cri-o://bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159" gracePeriod=2 Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.488886 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.871718 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.992160 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-utilities\") pod \"2efacc1c-1a99-4105-8034-b7554d1013f9\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.992314 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-catalog-content\") pod \"2efacc1c-1a99-4105-8034-b7554d1013f9\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.992381 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25l9p\" (UniqueName: \"kubernetes.io/projected/2efacc1c-1a99-4105-8034-b7554d1013f9-kube-api-access-25l9p\") pod \"2efacc1c-1a99-4105-8034-b7554d1013f9\" (UID: \"2efacc1c-1a99-4105-8034-b7554d1013f9\") " Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.995970 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-utilities" (OuterVolumeSpecName: "utilities") pod "2efacc1c-1a99-4105-8034-b7554d1013f9" (UID: "2efacc1c-1a99-4105-8034-b7554d1013f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:15:52 crc kubenswrapper[4894]: I1208 15:15:52.997623 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2efacc1c-1a99-4105-8034-b7554d1013f9-kube-api-access-25l9p" (OuterVolumeSpecName: "kube-api-access-25l9p") pod "2efacc1c-1a99-4105-8034-b7554d1013f9" (UID: "2efacc1c-1a99-4105-8034-b7554d1013f9"). InnerVolumeSpecName "kube-api-access-25l9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.057254 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2efacc1c-1a99-4105-8034-b7554d1013f9" (UID: "2efacc1c-1a99-4105-8034-b7554d1013f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.094323 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.094355 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2efacc1c-1a99-4105-8034-b7554d1013f9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.094385 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25l9p\" (UniqueName: \"kubernetes.io/projected/2efacc1c-1a99-4105-8034-b7554d1013f9-kube-api-access-25l9p\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.451743 4894 generic.go:334] "Generic (PLEG): container finished" podID="dff601df-c402-489d-b878-a8aaa7a32909" containerID="f2942d4e6a538eb18f3c3cd5b55f94c975ac36a5c8a5273cb4bf716bff7ddfab" exitCode=0 Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.451864 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" event={"ID":"dff601df-c402-489d-b878-a8aaa7a32909","Type":"ContainerDied","Data":"f2942d4e6a538eb18f3c3cd5b55f94c975ac36a5c8a5273cb4bf716bff7ddfab"} Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.455166 4894 generic.go:334] "Generic (PLEG): container finished" podID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerID="bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159" exitCode=0 Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.456062 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrdms" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.456546 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrdms" event={"ID":"2efacc1c-1a99-4105-8034-b7554d1013f9","Type":"ContainerDied","Data":"bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159"} Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.456571 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrdms" event={"ID":"2efacc1c-1a99-4105-8034-b7554d1013f9","Type":"ContainerDied","Data":"16a410f21061061dcc56a4039724856e49fa475c811a155f774f7d9e973c3a8c"} Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.456588 4894 scope.go:117] "RemoveContainer" containerID="bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.479358 4894 scope.go:117] "RemoveContainer" containerID="c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.507831 4894 scope.go:117] "RemoveContainer" containerID="5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.512684 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jrdms"] Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.525212 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jrdms"] Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.570121 4894 scope.go:117] "RemoveContainer" containerID="bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159" Dec 08 15:15:53 crc kubenswrapper[4894]: E1208 15:15:53.570690 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159\": container with ID starting with bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159 not found: ID does not exist" containerID="bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.570749 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159"} err="failed to get container status \"bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159\": rpc error: code = NotFound desc = could not find container \"bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159\": container with ID starting with bab73d8ffa45a45eeada1e9cf70b7aa8b5a3177301759b41cabe3b7b3b87f159 not found: ID does not exist" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.570781 4894 scope.go:117] "RemoveContainer" containerID="c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0" Dec 08 15:15:53 crc kubenswrapper[4894]: E1208 15:15:53.571321 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0\": container with ID starting with c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0 not found: ID does not exist" containerID="c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.571350 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0"} err="failed to get container status \"c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0\": rpc error: code = NotFound desc = could not find container \"c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0\": container with ID starting with c84d43a1461dc06011c7e5d72a7e5bf2d1c7d2452b3a479e4ec13961652ba3a0 not found: ID does not exist" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.571368 4894 scope.go:117] "RemoveContainer" containerID="5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764" Dec 08 15:15:53 crc kubenswrapper[4894]: E1208 15:15:53.571836 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764\": container with ID starting with 5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764 not found: ID does not exist" containerID="5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764" Dec 08 15:15:53 crc kubenswrapper[4894]: I1208 15:15:53.571860 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764"} err="failed to get container status \"5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764\": rpc error: code = NotFound desc = could not find container \"5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764\": container with ID starting with 5e4e47b638ba4859833e20c425c8f308c2cbd505ecaa6bf09a2aab4817ab5764 not found: ID does not exist" Dec 08 15:15:54 crc kubenswrapper[4894]: I1208 15:15:54.833986 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.030049 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8287\" (UniqueName: \"kubernetes.io/projected/dff601df-c402-489d-b878-a8aaa7a32909-kube-api-access-p8287\") pod \"dff601df-c402-489d-b878-a8aaa7a32909\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.030231 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-inventory\") pod \"dff601df-c402-489d-b878-a8aaa7a32909\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.030310 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-ssh-key\") pod \"dff601df-c402-489d-b878-a8aaa7a32909\" (UID: \"dff601df-c402-489d-b878-a8aaa7a32909\") " Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.036743 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dff601df-c402-489d-b878-a8aaa7a32909-kube-api-access-p8287" (OuterVolumeSpecName: "kube-api-access-p8287") pod "dff601df-c402-489d-b878-a8aaa7a32909" (UID: "dff601df-c402-489d-b878-a8aaa7a32909"). InnerVolumeSpecName "kube-api-access-p8287". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.065404 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-inventory" (OuterVolumeSpecName: "inventory") pod "dff601df-c402-489d-b878-a8aaa7a32909" (UID: "dff601df-c402-489d-b878-a8aaa7a32909"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.074009 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dff601df-c402-489d-b878-a8aaa7a32909" (UID: "dff601df-c402-489d-b878-a8aaa7a32909"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.132996 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8287\" (UniqueName: \"kubernetes.io/projected/dff601df-c402-489d-b878-a8aaa7a32909-kube-api-access-p8287\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.133037 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.133051 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dff601df-c402-489d-b878-a8aaa7a32909-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.208967 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" path="/var/lib/kubelet/pods/2efacc1c-1a99-4105-8034-b7554d1013f9/volumes" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.476876 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" event={"ID":"dff601df-c402-489d-b878-a8aaa7a32909","Type":"ContainerDied","Data":"b84d47477213387bbb99b1e4a9a2e81e3f0e3d6bd61e72b7ff7b123be5b72d0a"} Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.477201 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b84d47477213387bbb99b1e4a9a2e81e3f0e3d6bd61e72b7ff7b123be5b72d0a" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.476941 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vcccf" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.581336 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx"] Dec 08 15:15:55 crc kubenswrapper[4894]: E1208 15:15:55.581860 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="extract-content" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.581878 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="extract-content" Dec 08 15:15:55 crc kubenswrapper[4894]: E1208 15:15:55.581898 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="registry-server" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.581905 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="registry-server" Dec 08 15:15:55 crc kubenswrapper[4894]: E1208 15:15:55.581930 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff601df-c402-489d-b878-a8aaa7a32909" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.581940 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff601df-c402-489d-b878-a8aaa7a32909" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:55 crc kubenswrapper[4894]: E1208 15:15:55.581960 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="extract-utilities" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.581968 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="extract-utilities" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.582207 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="dff601df-c402-489d-b878-a8aaa7a32909" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.582238 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2efacc1c-1a99-4105-8034-b7554d1013f9" containerName="registry-server" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.583002 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.585101 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.585138 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.585407 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.586005 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.593298 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx"] Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.643024 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-689nf\" (UniqueName: \"kubernetes.io/projected/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-kube-api-access-689nf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.643251 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.643420 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.745769 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.745889 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-689nf\" (UniqueName: \"kubernetes.io/projected/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-kube-api-access-689nf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.745976 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.750298 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.750844 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.762695 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-689nf\" (UniqueName: \"kubernetes.io/projected/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-kube-api-access-689nf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:55 crc kubenswrapper[4894]: I1208 15:15:55.916247 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.197376 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:15:56 crc kubenswrapper[4894]: E1208 15:15:56.198101 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.499793 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx"] Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.776609 4894 scope.go:117] "RemoveContainer" containerID="52b30224bcfff18b6d78842f366c717a5842da23c245ee7ca9c0a6f2ff55dbfb" Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.803776 4894 scope.go:117] "RemoveContainer" containerID="8e0eff55444972e050b2d1c575a3a5d7c6de042add02fb2c3953f8fdcdfef3f5" Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.877966 4894 scope.go:117] "RemoveContainer" containerID="3daa8390366acee3732b5671a4160901220b017445b783e7d47984580a880039" Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.936598 4894 scope.go:117] "RemoveContainer" containerID="ef31a8436b105850224ef08c14f32232599779e1eabc7a8b36d5df4cf16c35a2" Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.940951 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nt99r"] Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.941179 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nt99r" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="registry-server" containerID="cri-o://35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8" gracePeriod=2 Dec 08 15:15:56 crc kubenswrapper[4894]: I1208 15:15:56.979261 4894 scope.go:117] "RemoveContainer" containerID="be5cfa049be2274e6588909adc1b360ddc25f2e4106ac48e1f25de455dd749af" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.130798 4894 scope.go:117] "RemoveContainer" containerID="0db2de46b62ee56858e9dccbea460b5bdc4a012d009472831e81dd2f6aeba3d8" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.185243 4894 scope.go:117] "RemoveContainer" containerID="fb7f69ae479eaaa6a65aaf7d990283ba529dcea3cab4d0a8e611f0a100c8de01" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.213230 4894 scope.go:117] "RemoveContainer" containerID="55e524ca71c981b718fa4cb601f1fc6601348c28640edb9425ce62cb401a1f29" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.363598 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.381337 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-utilities\") pod \"98c017a5-0df0-4e05-a7f8-a2d150910c44\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.381620 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4slc\" (UniqueName: \"kubernetes.io/projected/98c017a5-0df0-4e05-a7f8-a2d150910c44-kube-api-access-n4slc\") pod \"98c017a5-0df0-4e05-a7f8-a2d150910c44\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.381678 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-catalog-content\") pod \"98c017a5-0df0-4e05-a7f8-a2d150910c44\" (UID: \"98c017a5-0df0-4e05-a7f8-a2d150910c44\") " Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.382298 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-utilities" (OuterVolumeSpecName: "utilities") pod "98c017a5-0df0-4e05-a7f8-a2d150910c44" (UID: "98c017a5-0df0-4e05-a7f8-a2d150910c44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.382614 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.387171 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c017a5-0df0-4e05-a7f8-a2d150910c44-kube-api-access-n4slc" (OuterVolumeSpecName: "kube-api-access-n4slc") pod "98c017a5-0df0-4e05-a7f8-a2d150910c44" (UID: "98c017a5-0df0-4e05-a7f8-a2d150910c44"). InnerVolumeSpecName "kube-api-access-n4slc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.403110 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98c017a5-0df0-4e05-a7f8-a2d150910c44" (UID: "98c017a5-0df0-4e05-a7f8-a2d150910c44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.484318 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4slc\" (UniqueName: \"kubernetes.io/projected/98c017a5-0df0-4e05-a7f8-a2d150910c44-kube-api-access-n4slc\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.484347 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98c017a5-0df0-4e05-a7f8-a2d150910c44-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.494705 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" event={"ID":"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a","Type":"ContainerStarted","Data":"a5679eb6af1e0f0d784721987bbac399867c8bc8dda9e39e93b37082f641c86c"} Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.494750 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" event={"ID":"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a","Type":"ContainerStarted","Data":"9a9c1ba9d1e0e6239f267587389a8c5df4097c7fbe4912c61416c0b1cf435e19"} Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.498791 4894 generic.go:334] "Generic (PLEG): container finished" podID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerID="35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8" exitCode=0 Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.498849 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nt99r" event={"ID":"98c017a5-0df0-4e05-a7f8-a2d150910c44","Type":"ContainerDied","Data":"35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8"} Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.498891 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nt99r" event={"ID":"98c017a5-0df0-4e05-a7f8-a2d150910c44","Type":"ContainerDied","Data":"64a49d539fedfd41c2b6ea10d010f9f62c3db82764ee76a9dada83954fff1345"} Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.498919 4894 scope.go:117] "RemoveContainer" containerID="35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.499384 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nt99r" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.512284 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" podStartSLOduration=2.019644026 podStartE2EDuration="2.512268335s" podCreationTimestamp="2025-12-08 15:15:55 +0000 UTC" firstStartedPulling="2025-12-08 15:15:56.506018242 +0000 UTC m=+1777.606024357" lastFinishedPulling="2025-12-08 15:15:56.998642551 +0000 UTC m=+1778.098648666" observedRunningTime="2025-12-08 15:15:57.511983496 +0000 UTC m=+1778.611989631" watchObservedRunningTime="2025-12-08 15:15:57.512268335 +0000 UTC m=+1778.612274450" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.526755 4894 scope.go:117] "RemoveContainer" containerID="dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.542005 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nt99r"] Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.554953 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nt99r"] Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.556682 4894 scope.go:117] "RemoveContainer" containerID="ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.585323 4894 scope.go:117] "RemoveContainer" containerID="35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8" Dec 08 15:15:57 crc kubenswrapper[4894]: E1208 15:15:57.588140 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8\": container with ID starting with 35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8 not found: ID does not exist" containerID="35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.588182 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8"} err="failed to get container status \"35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8\": rpc error: code = NotFound desc = could not find container \"35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8\": container with ID starting with 35889cdda7c089fdc28e3b43c8aa3f80ebb36a17234e44f4a7baa82cd05fb7b8 not found: ID does not exist" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.588211 4894 scope.go:117] "RemoveContainer" containerID="dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928" Dec 08 15:15:57 crc kubenswrapper[4894]: E1208 15:15:57.590938 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928\": container with ID starting with dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928 not found: ID does not exist" containerID="dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.590975 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928"} err="failed to get container status \"dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928\": rpc error: code = NotFound desc = could not find container \"dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928\": container with ID starting with dfc73332f75b86c197a0cbda3f862517c76baca89a0ea948f616e84ff3b02928 not found: ID does not exist" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.590998 4894 scope.go:117] "RemoveContainer" containerID="ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1" Dec 08 15:15:57 crc kubenswrapper[4894]: E1208 15:15:57.591261 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1\": container with ID starting with ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1 not found: ID does not exist" containerID="ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1" Dec 08 15:15:57 crc kubenswrapper[4894]: I1208 15:15:57.591297 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1"} err="failed to get container status \"ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1\": rpc error: code = NotFound desc = could not find container \"ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1\": container with ID starting with ab8f5f0646075055e494932d6489c9d50bc0ac0e6e35c99aa3c47fa21f5c4fa1 not found: ID does not exist" Dec 08 15:15:59 crc kubenswrapper[4894]: I1208 15:15:59.207559 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" path="/var/lib/kubelet/pods/98c017a5-0df0-4e05-a7f8-a2d150910c44/volumes" Dec 08 15:16:11 crc kubenswrapper[4894]: I1208 15:16:11.197094 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:16:11 crc kubenswrapper[4894]: E1208 15:16:11.197780 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:16:12 crc kubenswrapper[4894]: I1208 15:16:12.038338 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjlmx"] Dec 08 15:16:12 crc kubenswrapper[4894]: I1208 15:16:12.046991 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-pjlmx"] Dec 08 15:16:13 crc kubenswrapper[4894]: I1208 15:16:13.207643 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c418af0-bc4d-4fe4-880a-804fd7412ec8" path="/var/lib/kubelet/pods/6c418af0-bc4d-4fe4-880a-804fd7412ec8/volumes" Dec 08 15:16:22 crc kubenswrapper[4894]: I1208 15:16:22.197525 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:16:22 crc kubenswrapper[4894]: E1208 15:16:22.198444 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:16:35 crc kubenswrapper[4894]: I1208 15:16:35.199011 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:16:35 crc kubenswrapper[4894]: E1208 15:16:35.201700 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:16:36 crc kubenswrapper[4894]: I1208 15:16:36.052017 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-mjdxz"] Dec 08 15:16:36 crc kubenswrapper[4894]: I1208 15:16:36.059251 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-mjdxz"] Dec 08 15:16:37 crc kubenswrapper[4894]: I1208 15:16:37.207782 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06425613-4821-4a3d-9507-f6176518464b" path="/var/lib/kubelet/pods/06425613-4821-4a3d-9507-f6176518464b/volumes" Dec 08 15:16:38 crc kubenswrapper[4894]: I1208 15:16:38.047237 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2b5qb"] Dec 08 15:16:38 crc kubenswrapper[4894]: I1208 15:16:38.065589 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2b5qb"] Dec 08 15:16:39 crc kubenswrapper[4894]: I1208 15:16:39.208786 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d73ecea1-981c-4b0a-91f1-ace506d48483" path="/var/lib/kubelet/pods/d73ecea1-981c-4b0a-91f1-ace506d48483/volumes" Dec 08 15:16:47 crc kubenswrapper[4894]: I1208 15:16:47.987624 4894 generic.go:334] "Generic (PLEG): container finished" podID="0f36f23c-cd37-4fe4-b851-6f1aa2c4017a" containerID="a5679eb6af1e0f0d784721987bbac399867c8bc8dda9e39e93b37082f641c86c" exitCode=0 Dec 08 15:16:47 crc kubenswrapper[4894]: I1208 15:16:47.987730 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" event={"ID":"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a","Type":"ContainerDied","Data":"a5679eb6af1e0f0d784721987bbac399867c8bc8dda9e39e93b37082f641c86c"} Dec 08 15:16:48 crc kubenswrapper[4894]: I1208 15:16:48.199411 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:16:48 crc kubenswrapper[4894]: E1208 15:16:48.199646 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.382870 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.449151 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-689nf\" (UniqueName: \"kubernetes.io/projected/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-kube-api-access-689nf\") pod \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.449305 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-inventory\") pod \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.449374 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-ssh-key\") pod \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\" (UID: \"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a\") " Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.454986 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-kube-api-access-689nf" (OuterVolumeSpecName: "kube-api-access-689nf") pod "0f36f23c-cd37-4fe4-b851-6f1aa2c4017a" (UID: "0f36f23c-cd37-4fe4-b851-6f1aa2c4017a"). InnerVolumeSpecName "kube-api-access-689nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.479355 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0f36f23c-cd37-4fe4-b851-6f1aa2c4017a" (UID: "0f36f23c-cd37-4fe4-b851-6f1aa2c4017a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.488962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-inventory" (OuterVolumeSpecName: "inventory") pod "0f36f23c-cd37-4fe4-b851-6f1aa2c4017a" (UID: "0f36f23c-cd37-4fe4-b851-6f1aa2c4017a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.551983 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-689nf\" (UniqueName: \"kubernetes.io/projected/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-kube-api-access-689nf\") on node \"crc\" DevicePath \"\"" Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.552295 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:16:49 crc kubenswrapper[4894]: I1208 15:16:49.552305 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f36f23c-cd37-4fe4-b851-6f1aa2c4017a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.005561 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" event={"ID":"0f36f23c-cd37-4fe4-b851-6f1aa2c4017a","Type":"ContainerDied","Data":"9a9c1ba9d1e0e6239f267587389a8c5df4097c7fbe4912c61416c0b1cf435e19"} Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.005605 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a9c1ba9d1e0e6239f267587389a8c5df4097c7fbe4912c61416c0b1cf435e19" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.005662 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.088914 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-m87rg"] Dec 08 15:16:50 crc kubenswrapper[4894]: E1208 15:16:50.089577 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f36f23c-cd37-4fe4-b851-6f1aa2c4017a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.089650 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f36f23c-cd37-4fe4-b851-6f1aa2c4017a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:16:50 crc kubenswrapper[4894]: E1208 15:16:50.089708 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="registry-server" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.089789 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="registry-server" Dec 08 15:16:50 crc kubenswrapper[4894]: E1208 15:16:50.089893 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="extract-content" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.090051 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="extract-content" Dec 08 15:16:50 crc kubenswrapper[4894]: E1208 15:16:50.090109 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="extract-utilities" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.090163 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="extract-utilities" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.090389 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c017a5-0df0-4e05-a7f8-a2d150910c44" containerName="registry-server" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.090465 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f36f23c-cd37-4fe4-b851-6f1aa2c4017a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.091139 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.100903 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.101617 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.115351 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-m87rg"] Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.116294 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.117148 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.160625 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qjfg\" (UniqueName: \"kubernetes.io/projected/b5b86fe7-04aa-4538-88d1-d6dd42673b07-kube-api-access-2qjfg\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.160690 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.160712 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.264047 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qjfg\" (UniqueName: \"kubernetes.io/projected/b5b86fe7-04aa-4538-88d1-d6dd42673b07-kube-api-access-2qjfg\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.264128 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.264165 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.274104 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.288198 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.289000 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qjfg\" (UniqueName: \"kubernetes.io/projected/b5b86fe7-04aa-4538-88d1-d6dd42673b07-kube-api-access-2qjfg\") pod \"ssh-known-hosts-edpm-deployment-m87rg\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.411771 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:50 crc kubenswrapper[4894]: I1208 15:16:50.909208 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-m87rg"] Dec 08 15:16:51 crc kubenswrapper[4894]: I1208 15:16:51.015007 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" event={"ID":"b5b86fe7-04aa-4538-88d1-d6dd42673b07","Type":"ContainerStarted","Data":"6559f617a275fe84475758e1c7ffdaaca7d5a7a8ae2a5d73eaa995b94f40a7af"} Dec 08 15:16:52 crc kubenswrapper[4894]: I1208 15:16:52.029059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" event={"ID":"b5b86fe7-04aa-4538-88d1-d6dd42673b07","Type":"ContainerStarted","Data":"1c66b0b2fe9e2371b7236ace8017e658c1f223097be5bfc716f28a801da15706"} Dec 08 15:16:52 crc kubenswrapper[4894]: I1208 15:16:52.051984 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" podStartSLOduration=1.646668988 podStartE2EDuration="2.051966446s" podCreationTimestamp="2025-12-08 15:16:50 +0000 UTC" firstStartedPulling="2025-12-08 15:16:50.918040223 +0000 UTC m=+1832.018046338" lastFinishedPulling="2025-12-08 15:16:51.323337681 +0000 UTC m=+1832.423343796" observedRunningTime="2025-12-08 15:16:52.045596057 +0000 UTC m=+1833.145602172" watchObservedRunningTime="2025-12-08 15:16:52.051966446 +0000 UTC m=+1833.151972561" Dec 08 15:16:57 crc kubenswrapper[4894]: I1208 15:16:57.437967 4894 scope.go:117] "RemoveContainer" containerID="6348e27b10880292ffc95b16fe26fc6300ab2e14fa8daa23a5b55a13e3a76e2b" Dec 08 15:16:57 crc kubenswrapper[4894]: I1208 15:16:57.493163 4894 scope.go:117] "RemoveContainer" containerID="5c93a8fb060ccaf2f3a853961218545d48cece43a590aea4fc345486ed51c89f" Dec 08 15:16:57 crc kubenswrapper[4894]: I1208 15:16:57.533523 4894 scope.go:117] "RemoveContainer" containerID="f82f44abbb6f6215eca56ec63cd083e5e05ee38f02998d93101a0c3443746003" Dec 08 15:16:58 crc kubenswrapper[4894]: I1208 15:16:58.092403 4894 generic.go:334] "Generic (PLEG): container finished" podID="b5b86fe7-04aa-4538-88d1-d6dd42673b07" containerID="1c66b0b2fe9e2371b7236ace8017e658c1f223097be5bfc716f28a801da15706" exitCode=0 Dec 08 15:16:58 crc kubenswrapper[4894]: I1208 15:16:58.092447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" event={"ID":"b5b86fe7-04aa-4538-88d1-d6dd42673b07","Type":"ContainerDied","Data":"1c66b0b2fe9e2371b7236ace8017e658c1f223097be5bfc716f28a801da15706"} Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.474492 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.633378 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-inventory-0\") pod \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.633720 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qjfg\" (UniqueName: \"kubernetes.io/projected/b5b86fe7-04aa-4538-88d1-d6dd42673b07-kube-api-access-2qjfg\") pod \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.633878 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-ssh-key-openstack-edpm-ipam\") pod \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\" (UID: \"b5b86fe7-04aa-4538-88d1-d6dd42673b07\") " Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.660318 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b5b86fe7-04aa-4538-88d1-d6dd42673b07" (UID: "b5b86fe7-04aa-4538-88d1-d6dd42673b07"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.661484 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b86fe7-04aa-4538-88d1-d6dd42673b07-kube-api-access-2qjfg" (OuterVolumeSpecName: "kube-api-access-2qjfg") pod "b5b86fe7-04aa-4538-88d1-d6dd42673b07" (UID: "b5b86fe7-04aa-4538-88d1-d6dd42673b07"). InnerVolumeSpecName "kube-api-access-2qjfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.663046 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b5b86fe7-04aa-4538-88d1-d6dd42673b07" (UID: "b5b86fe7-04aa-4538-88d1-d6dd42673b07"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.736496 4894 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.736683 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qjfg\" (UniqueName: \"kubernetes.io/projected/b5b86fe7-04aa-4538-88d1-d6dd42673b07-kube-api-access-2qjfg\") on node \"crc\" DevicePath \"\"" Dec 08 15:16:59 crc kubenswrapper[4894]: I1208 15:16:59.736702 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b5b86fe7-04aa-4538-88d1-d6dd42673b07-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.109201 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" event={"ID":"b5b86fe7-04aa-4538-88d1-d6dd42673b07","Type":"ContainerDied","Data":"6559f617a275fe84475758e1c7ffdaaca7d5a7a8ae2a5d73eaa995b94f40a7af"} Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.109242 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6559f617a275fe84475758e1c7ffdaaca7d5a7a8ae2a5d73eaa995b94f40a7af" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.109251 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-m87rg" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.193450 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6"] Dec 08 15:17:00 crc kubenswrapper[4894]: E1208 15:17:00.194073 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b86fe7-04aa-4538-88d1-d6dd42673b07" containerName="ssh-known-hosts-edpm-deployment" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.194093 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b86fe7-04aa-4538-88d1-d6dd42673b07" containerName="ssh-known-hosts-edpm-deployment" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.194309 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b86fe7-04aa-4538-88d1-d6dd42673b07" containerName="ssh-known-hosts-edpm-deployment" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.195048 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.196622 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:17:00 crc kubenswrapper[4894]: E1208 15:17:00.196876 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.197047 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.197399 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.197691 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.197883 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.230453 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6"] Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.352862 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kclw5\" (UniqueName: \"kubernetes.io/projected/f840f001-3515-40c7-bc6e-6434ee8df42a-kube-api-access-kclw5\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.353044 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.353528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.455327 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.455406 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kclw5\" (UniqueName: \"kubernetes.io/projected/f840f001-3515-40c7-bc6e-6434ee8df42a-kube-api-access-kclw5\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.455461 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.462829 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.463472 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.474255 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kclw5\" (UniqueName: \"kubernetes.io/projected/f840f001-3515-40c7-bc6e-6434ee8df42a-kube-api-access-kclw5\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lxvp6\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:00 crc kubenswrapper[4894]: I1208 15:17:00.521116 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:01 crc kubenswrapper[4894]: I1208 15:17:01.023877 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6"] Dec 08 15:17:01 crc kubenswrapper[4894]: I1208 15:17:01.121191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" event={"ID":"f840f001-3515-40c7-bc6e-6434ee8df42a","Type":"ContainerStarted","Data":"6882e7c35c716c8c1e96c4849236ce8a11fabf0760890e3260efa7f7d760bbaf"} Dec 08 15:17:02 crc kubenswrapper[4894]: I1208 15:17:02.129284 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" event={"ID":"f840f001-3515-40c7-bc6e-6434ee8df42a","Type":"ContainerStarted","Data":"fc7b5c3e513979ed809e94b24baa48288497f871b31ca860e6e11ad2e720a26a"} Dec 08 15:17:02 crc kubenswrapper[4894]: I1208 15:17:02.152875 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" podStartSLOduration=1.725418444 podStartE2EDuration="2.152856294s" podCreationTimestamp="2025-12-08 15:17:00 +0000 UTC" firstStartedPulling="2025-12-08 15:17:01.027358934 +0000 UTC m=+1842.127365059" lastFinishedPulling="2025-12-08 15:17:01.454796794 +0000 UTC m=+1842.554802909" observedRunningTime="2025-12-08 15:17:02.142031436 +0000 UTC m=+1843.242037551" watchObservedRunningTime="2025-12-08 15:17:02.152856294 +0000 UTC m=+1843.252862409" Dec 08 15:17:10 crc kubenswrapper[4894]: I1208 15:17:10.208904 4894 generic.go:334] "Generic (PLEG): container finished" podID="f840f001-3515-40c7-bc6e-6434ee8df42a" containerID="fc7b5c3e513979ed809e94b24baa48288497f871b31ca860e6e11ad2e720a26a" exitCode=0 Dec 08 15:17:10 crc kubenswrapper[4894]: I1208 15:17:10.208985 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" event={"ID":"f840f001-3515-40c7-bc6e-6434ee8df42a","Type":"ContainerDied","Data":"fc7b5c3e513979ed809e94b24baa48288497f871b31ca860e6e11ad2e720a26a"} Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.604838 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.775505 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-inventory\") pod \"f840f001-3515-40c7-bc6e-6434ee8df42a\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.776005 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-ssh-key\") pod \"f840f001-3515-40c7-bc6e-6434ee8df42a\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.776189 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kclw5\" (UniqueName: \"kubernetes.io/projected/f840f001-3515-40c7-bc6e-6434ee8df42a-kube-api-access-kclw5\") pod \"f840f001-3515-40c7-bc6e-6434ee8df42a\" (UID: \"f840f001-3515-40c7-bc6e-6434ee8df42a\") " Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.798192 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f840f001-3515-40c7-bc6e-6434ee8df42a-kube-api-access-kclw5" (OuterVolumeSpecName: "kube-api-access-kclw5") pod "f840f001-3515-40c7-bc6e-6434ee8df42a" (UID: "f840f001-3515-40c7-bc6e-6434ee8df42a"). InnerVolumeSpecName "kube-api-access-kclw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.807836 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f840f001-3515-40c7-bc6e-6434ee8df42a" (UID: "f840f001-3515-40c7-bc6e-6434ee8df42a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.809486 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-inventory" (OuterVolumeSpecName: "inventory") pod "f840f001-3515-40c7-bc6e-6434ee8df42a" (UID: "f840f001-3515-40c7-bc6e-6434ee8df42a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.877949 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kclw5\" (UniqueName: \"kubernetes.io/projected/f840f001-3515-40c7-bc6e-6434ee8df42a-kube-api-access-kclw5\") on node \"crc\" DevicePath \"\"" Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.877981 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:17:11 crc kubenswrapper[4894]: I1208 15:17:11.877992 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f840f001-3515-40c7-bc6e-6434ee8df42a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.197099 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.226194 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.226194 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lxvp6" event={"ID":"f840f001-3515-40c7-bc6e-6434ee8df42a","Type":"ContainerDied","Data":"6882e7c35c716c8c1e96c4849236ce8a11fabf0760890e3260efa7f7d760bbaf"} Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.226320 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6882e7c35c716c8c1e96c4849236ce8a11fabf0760890e3260efa7f7d760bbaf" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.292888 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv"] Dec 08 15:17:12 crc kubenswrapper[4894]: E1208 15:17:12.293545 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f840f001-3515-40c7-bc6e-6434ee8df42a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.293567 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f840f001-3515-40c7-bc6e-6434ee8df42a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.293749 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f840f001-3515-40c7-bc6e-6434ee8df42a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.294400 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.296526 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.296806 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.297026 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.297110 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.302730 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv"] Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.489914 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.489993 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.490025 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk6gk\" (UniqueName: \"kubernetes.io/projected/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-kube-api-access-mk6gk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.591345 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.591413 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk6gk\" (UniqueName: \"kubernetes.io/projected/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-kube-api-access-mk6gk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.591566 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.597117 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.597124 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.616582 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk6gk\" (UniqueName: \"kubernetes.io/projected/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-kube-api-access-mk6gk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:12 crc kubenswrapper[4894]: I1208 15:17:12.634230 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:13 crc kubenswrapper[4894]: I1208 15:17:13.135006 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv"] Dec 08 15:17:13 crc kubenswrapper[4894]: I1208 15:17:13.239711 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"f23f6da90e9a9fd343cea89c0d4e56edbcbde0c11dbbfcf77ae9c0fffc82c8c3"} Dec 08 15:17:13 crc kubenswrapper[4894]: I1208 15:17:13.242106 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" event={"ID":"fae74a7d-767f-47a9-8f3c-1337c1be1d3e","Type":"ContainerStarted","Data":"75a9f6b3202b91c0e1e29c9fef8f27bed349082a6566e08f550ad1ff6d134005"} Dec 08 15:17:14 crc kubenswrapper[4894]: I1208 15:17:14.251703 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" event={"ID":"fae74a7d-767f-47a9-8f3c-1337c1be1d3e","Type":"ContainerStarted","Data":"9ccc643e352287a229f4cb6ff7caebc08efeea4ac109f00bd87672962693f6fa"} Dec 08 15:17:14 crc kubenswrapper[4894]: I1208 15:17:14.276542 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" podStartSLOduration=1.867539895 podStartE2EDuration="2.276519249s" podCreationTimestamp="2025-12-08 15:17:12 +0000 UTC" firstStartedPulling="2025-12-08 15:17:13.14305584 +0000 UTC m=+1854.243061955" lastFinishedPulling="2025-12-08 15:17:13.552035194 +0000 UTC m=+1854.652041309" observedRunningTime="2025-12-08 15:17:14.272331899 +0000 UTC m=+1855.372338024" watchObservedRunningTime="2025-12-08 15:17:14.276519249 +0000 UTC m=+1855.376525374" Dec 08 15:17:20 crc kubenswrapper[4894]: I1208 15:17:20.045878 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-lkqq6"] Dec 08 15:17:20 crc kubenswrapper[4894]: I1208 15:17:20.057193 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-lkqq6"] Dec 08 15:17:21 crc kubenswrapper[4894]: I1208 15:17:21.207560 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f846ca80-7b63-4a22-b639-768158cd85c3" path="/var/lib/kubelet/pods/f846ca80-7b63-4a22-b639-768158cd85c3/volumes" Dec 08 15:17:22 crc kubenswrapper[4894]: I1208 15:17:22.325358 4894 generic.go:334] "Generic (PLEG): container finished" podID="fae74a7d-767f-47a9-8f3c-1337c1be1d3e" containerID="9ccc643e352287a229f4cb6ff7caebc08efeea4ac109f00bd87672962693f6fa" exitCode=0 Dec 08 15:17:22 crc kubenswrapper[4894]: I1208 15:17:22.325424 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" event={"ID":"fae74a7d-767f-47a9-8f3c-1337c1be1d3e","Type":"ContainerDied","Data":"9ccc643e352287a229f4cb6ff7caebc08efeea4ac109f00bd87672962693f6fa"} Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.704878 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.800414 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-ssh-key\") pod \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.800609 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-inventory\") pod \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.800730 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk6gk\" (UniqueName: \"kubernetes.io/projected/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-kube-api-access-mk6gk\") pod \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\" (UID: \"fae74a7d-767f-47a9-8f3c-1337c1be1d3e\") " Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.806305 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-kube-api-access-mk6gk" (OuterVolumeSpecName: "kube-api-access-mk6gk") pod "fae74a7d-767f-47a9-8f3c-1337c1be1d3e" (UID: "fae74a7d-767f-47a9-8f3c-1337c1be1d3e"). InnerVolumeSpecName "kube-api-access-mk6gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.829898 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-inventory" (OuterVolumeSpecName: "inventory") pod "fae74a7d-767f-47a9-8f3c-1337c1be1d3e" (UID: "fae74a7d-767f-47a9-8f3c-1337c1be1d3e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.831348 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fae74a7d-767f-47a9-8f3c-1337c1be1d3e" (UID: "fae74a7d-767f-47a9-8f3c-1337c1be1d3e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.903263 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk6gk\" (UniqueName: \"kubernetes.io/projected/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-kube-api-access-mk6gk\") on node \"crc\" DevicePath \"\"" Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.903301 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:17:23 crc kubenswrapper[4894]: I1208 15:17:23.903319 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fae74a7d-767f-47a9-8f3c-1337c1be1d3e-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.343191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" event={"ID":"fae74a7d-767f-47a9-8f3c-1337c1be1d3e","Type":"ContainerDied","Data":"75a9f6b3202b91c0e1e29c9fef8f27bed349082a6566e08f550ad1ff6d134005"} Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.343239 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75a9f6b3202b91c0e1e29c9fef8f27bed349082a6566e08f550ad1ff6d134005" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.343308 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.425233 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v"] Dec 08 15:17:24 crc kubenswrapper[4894]: E1208 15:17:24.425620 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fae74a7d-767f-47a9-8f3c-1337c1be1d3e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.425637 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fae74a7d-767f-47a9-8f3c-1337c1be1d3e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.425843 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fae74a7d-767f-47a9-8f3c-1337c1be1d3e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.426449 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.428718 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.428761 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.428720 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.428847 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.428941 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.432047 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.432095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.432378 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.440664 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v"] Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.519471 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.519516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.519543 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.519566 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.519604 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.519810 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.519953 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.520115 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.520160 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.520209 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhw58\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-kube-api-access-zhw58\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.520333 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.520380 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.520425 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.520562 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622214 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622318 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622513 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622588 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhw58\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-kube-api-access-zhw58\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622691 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622734 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.622858 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.623046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.623093 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.623134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.623170 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.623189 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.626634 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.626794 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.627320 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.627436 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.627877 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.627884 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.629602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.629853 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.630946 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.631550 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.631995 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.636366 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.640556 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.641694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhw58\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-kube-api-access-zhw58\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:24 crc kubenswrapper[4894]: I1208 15:17:24.744041 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:17:25 crc kubenswrapper[4894]: I1208 15:17:25.280462 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v"] Dec 08 15:17:25 crc kubenswrapper[4894]: I1208 15:17:25.285866 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:17:25 crc kubenswrapper[4894]: I1208 15:17:25.355107 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" event={"ID":"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa","Type":"ContainerStarted","Data":"7babaf445a3d827372580a7e693463964088bdae7126982f6c2b8b8cbafaffc5"} Dec 08 15:17:26 crc kubenswrapper[4894]: I1208 15:17:26.368298 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" event={"ID":"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa","Type":"ContainerStarted","Data":"4ae9fb47fc97b754c04c00a75b146c2bc20d8f6eb7b7ba89244c1f3e651e0111"} Dec 08 15:17:26 crc kubenswrapper[4894]: I1208 15:17:26.398291 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" podStartSLOduration=1.994142191 podStartE2EDuration="2.398265253s" podCreationTimestamp="2025-12-08 15:17:24 +0000 UTC" firstStartedPulling="2025-12-08 15:17:25.285624955 +0000 UTC m=+1866.385631080" lastFinishedPulling="2025-12-08 15:17:25.689748017 +0000 UTC m=+1866.789754142" observedRunningTime="2025-12-08 15:17:26.387076923 +0000 UTC m=+1867.487083038" watchObservedRunningTime="2025-12-08 15:17:26.398265253 +0000 UTC m=+1867.498271368" Dec 08 15:17:57 crc kubenswrapper[4894]: I1208 15:17:57.658474 4894 scope.go:117] "RemoveContainer" containerID="9b21b545fd572ccb99efe6565300d9a53548901f69ecdf10bf88f6535ed32a37" Dec 08 15:18:02 crc kubenswrapper[4894]: I1208 15:18:02.681810 4894 generic.go:334] "Generic (PLEG): container finished" podID="d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" containerID="4ae9fb47fc97b754c04c00a75b146c2bc20d8f6eb7b7ba89244c1f3e651e0111" exitCode=0 Dec 08 15:18:02 crc kubenswrapper[4894]: I1208 15:18:02.681853 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" event={"ID":"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa","Type":"ContainerDied","Data":"4ae9fb47fc97b754c04c00a75b146c2bc20d8f6eb7b7ba89244c1f3e651e0111"} Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.090269 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191087 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191162 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191200 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-neutron-metadata-combined-ca-bundle\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191218 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-nova-combined-ca-bundle\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191298 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191326 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191351 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-bootstrap-combined-ca-bundle\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191375 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-inventory\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191446 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ovn-combined-ca-bundle\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191482 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ssh-key\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191520 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhw58\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-kube-api-access-zhw58\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191541 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-libvirt-combined-ca-bundle\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191606 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-repo-setup-combined-ca-bundle\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.191657 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-telemetry-combined-ca-bundle\") pod \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\" (UID: \"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa\") " Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.198346 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.199246 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.199926 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.200514 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.200603 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.200654 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.201161 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.202082 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.202560 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.202708 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-kube-api-access-zhw58" (OuterVolumeSpecName: "kube-api-access-zhw58") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "kube-api-access-zhw58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.203351 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.203948 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.223344 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.224175 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-inventory" (OuterVolumeSpecName: "inventory") pod "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" (UID: "d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293534 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293578 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293589 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293598 4894 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293611 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293624 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293635 4894 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293647 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293660 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293672 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293682 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhw58\" (UniqueName: \"kubernetes.io/projected/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-kube-api-access-zhw58\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293691 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293700 4894 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.293709 4894 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.702478 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" event={"ID":"d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa","Type":"ContainerDied","Data":"7babaf445a3d827372580a7e693463964088bdae7126982f6c2b8b8cbafaffc5"} Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.703037 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7babaf445a3d827372580a7e693463964088bdae7126982f6c2b8b8cbafaffc5" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.702531 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.795077 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv"] Dec 08 15:18:04 crc kubenswrapper[4894]: E1208 15:18:04.795444 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.795462 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.795673 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.796279 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.797993 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.799391 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.799402 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.800265 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.801564 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.850206 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv"] Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.904986 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.905051 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5c4m\" (UniqueName: \"kubernetes.io/projected/0ccc212f-4efb-4444-9156-27751b81562e-kube-api-access-c5c4m\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.905301 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.905355 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0ccc212f-4efb-4444-9156-27751b81562e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:04 crc kubenswrapper[4894]: I1208 15:18:04.905499 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.007388 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.007581 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0ccc212f-4efb-4444-9156-27751b81562e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.007810 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.007975 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.008059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5c4m\" (UniqueName: \"kubernetes.io/projected/0ccc212f-4efb-4444-9156-27751b81562e-kube-api-access-c5c4m\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.010036 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0ccc212f-4efb-4444-9156-27751b81562e-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.013450 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.013699 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.017174 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.032137 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5c4m\" (UniqueName: \"kubernetes.io/projected/0ccc212f-4efb-4444-9156-27751b81562e-kube-api-access-c5c4m\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fvbzv\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.118223 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:18:05 crc kubenswrapper[4894]: I1208 15:18:05.753317 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv"] Dec 08 15:18:06 crc kubenswrapper[4894]: I1208 15:18:06.721700 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" event={"ID":"0ccc212f-4efb-4444-9156-27751b81562e","Type":"ContainerStarted","Data":"c9052d1856bff12e8a9f41950fff9ead9b1ba91fa61bec9a9dd9c426ed1ca5f3"} Dec 08 15:18:06 crc kubenswrapper[4894]: I1208 15:18:06.722182 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" event={"ID":"0ccc212f-4efb-4444-9156-27751b81562e","Type":"ContainerStarted","Data":"b5f1cba46e03ce525097d1fcc4ee96ceb1e4fc36e2baf47baf2a662758c5b719"} Dec 08 15:18:06 crc kubenswrapper[4894]: I1208 15:18:06.743644 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" podStartSLOduration=2.300747532 podStartE2EDuration="2.743624584s" podCreationTimestamp="2025-12-08 15:18:04 +0000 UTC" firstStartedPulling="2025-12-08 15:18:05.766605238 +0000 UTC m=+1906.866611343" lastFinishedPulling="2025-12-08 15:18:06.20948228 +0000 UTC m=+1907.309488395" observedRunningTime="2025-12-08 15:18:06.73678894 +0000 UTC m=+1907.836795055" watchObservedRunningTime="2025-12-08 15:18:06.743624584 +0000 UTC m=+1907.843630699" Dec 08 15:19:07 crc kubenswrapper[4894]: I1208 15:19:07.278916 4894 generic.go:334] "Generic (PLEG): container finished" podID="0ccc212f-4efb-4444-9156-27751b81562e" containerID="c9052d1856bff12e8a9f41950fff9ead9b1ba91fa61bec9a9dd9c426ed1ca5f3" exitCode=0 Dec 08 15:19:07 crc kubenswrapper[4894]: I1208 15:19:07.279795 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" event={"ID":"0ccc212f-4efb-4444-9156-27751b81562e","Type":"ContainerDied","Data":"c9052d1856bff12e8a9f41950fff9ead9b1ba91fa61bec9a9dd9c426ed1ca5f3"} Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.663964 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.828247 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-inventory\") pod \"0ccc212f-4efb-4444-9156-27751b81562e\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.828780 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ssh-key\") pod \"0ccc212f-4efb-4444-9156-27751b81562e\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.828917 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0ccc212f-4efb-4444-9156-27751b81562e-ovncontroller-config-0\") pod \"0ccc212f-4efb-4444-9156-27751b81562e\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.828984 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5c4m\" (UniqueName: \"kubernetes.io/projected/0ccc212f-4efb-4444-9156-27751b81562e-kube-api-access-c5c4m\") pod \"0ccc212f-4efb-4444-9156-27751b81562e\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.829059 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ovn-combined-ca-bundle\") pod \"0ccc212f-4efb-4444-9156-27751b81562e\" (UID: \"0ccc212f-4efb-4444-9156-27751b81562e\") " Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.835183 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ccc212f-4efb-4444-9156-27751b81562e-kube-api-access-c5c4m" (OuterVolumeSpecName: "kube-api-access-c5c4m") pod "0ccc212f-4efb-4444-9156-27751b81562e" (UID: "0ccc212f-4efb-4444-9156-27751b81562e"). InnerVolumeSpecName "kube-api-access-c5c4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.835334 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0ccc212f-4efb-4444-9156-27751b81562e" (UID: "0ccc212f-4efb-4444-9156-27751b81562e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.856316 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0ccc212f-4efb-4444-9156-27751b81562e" (UID: "0ccc212f-4efb-4444-9156-27751b81562e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.856581 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ccc212f-4efb-4444-9156-27751b81562e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "0ccc212f-4efb-4444-9156-27751b81562e" (UID: "0ccc212f-4efb-4444-9156-27751b81562e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.858413 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-inventory" (OuterVolumeSpecName: "inventory") pod "0ccc212f-4efb-4444-9156-27751b81562e" (UID: "0ccc212f-4efb-4444-9156-27751b81562e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.930517 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.930549 4894 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0ccc212f-4efb-4444-9156-27751b81562e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.930560 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5c4m\" (UniqueName: \"kubernetes.io/projected/0ccc212f-4efb-4444-9156-27751b81562e-kube-api-access-c5c4m\") on node \"crc\" DevicePath \"\"" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.930570 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:19:08 crc kubenswrapper[4894]: I1208 15:19:08.930578 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ccc212f-4efb-4444-9156-27751b81562e-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.295589 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" event={"ID":"0ccc212f-4efb-4444-9156-27751b81562e","Type":"ContainerDied","Data":"b5f1cba46e03ce525097d1fcc4ee96ceb1e4fc36e2baf47baf2a662758c5b719"} Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.295968 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5f1cba46e03ce525097d1fcc4ee96ceb1e4fc36e2baf47baf2a662758c5b719" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.295651 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fvbzv" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.389131 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c"] Dec 08 15:19:09 crc kubenswrapper[4894]: E1208 15:19:09.389554 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ccc212f-4efb-4444-9156-27751b81562e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.389571 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ccc212f-4efb-4444-9156-27751b81562e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.389761 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ccc212f-4efb-4444-9156-27751b81562e" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.390406 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.392467 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.392556 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.394334 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.394419 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.394338 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.394905 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.398935 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c"] Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.541504 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.541591 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.541827 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpl4l\" (UniqueName: \"kubernetes.io/projected/35d2beec-8391-420e-a413-44fed1095880-kube-api-access-lpl4l\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.542032 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.542063 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.542120 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.644037 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.644097 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.644138 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.644186 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.644282 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.644364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpl4l\" (UniqueName: \"kubernetes.io/projected/35d2beec-8391-420e-a413-44fed1095880-kube-api-access-lpl4l\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.648955 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.649299 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.651250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.657036 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.660278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.662754 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpl4l\" (UniqueName: \"kubernetes.io/projected/35d2beec-8391-420e-a413-44fed1095880-kube-api-access-lpl4l\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:09 crc kubenswrapper[4894]: I1208 15:19:09.706194 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:19:10 crc kubenswrapper[4894]: I1208 15:19:10.189151 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c"] Dec 08 15:19:10 crc kubenswrapper[4894]: I1208 15:19:10.309477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" event={"ID":"35d2beec-8391-420e-a413-44fed1095880","Type":"ContainerStarted","Data":"f0db677c38ed7e23cd70d7b51df185d49e6d13d7267a9b2ed94a4eebd39ef442"} Dec 08 15:19:11 crc kubenswrapper[4894]: I1208 15:19:11.318205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" event={"ID":"35d2beec-8391-420e-a413-44fed1095880","Type":"ContainerStarted","Data":"04739fe11f725c8333fd4b763171f524ff5e72d4767524ca003dad8b33050c6a"} Dec 08 15:19:37 crc kubenswrapper[4894]: I1208 15:19:37.296716 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:19:37 crc kubenswrapper[4894]: I1208 15:19:37.297669 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:19:58 crc kubenswrapper[4894]: I1208 15:19:58.991500 4894 generic.go:334] "Generic (PLEG): container finished" podID="35d2beec-8391-420e-a413-44fed1095880" containerID="04739fe11f725c8333fd4b763171f524ff5e72d4767524ca003dad8b33050c6a" exitCode=0 Dec 08 15:19:58 crc kubenswrapper[4894]: I1208 15:19:58.991649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" event={"ID":"35d2beec-8391-420e-a413-44fed1095880","Type":"ContainerDied","Data":"04739fe11f725c8333fd4b763171f524ff5e72d4767524ca003dad8b33050c6a"} Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.405092 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.440642 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpl4l\" (UniqueName: \"kubernetes.io/projected/35d2beec-8391-420e-a413-44fed1095880-kube-api-access-lpl4l\") pod \"35d2beec-8391-420e-a413-44fed1095880\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.440852 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-ssh-key\") pod \"35d2beec-8391-420e-a413-44fed1095880\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.440965 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-inventory\") pod \"35d2beec-8391-420e-a413-44fed1095880\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.441144 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-ovn-metadata-agent-neutron-config-0\") pod \"35d2beec-8391-420e-a413-44fed1095880\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.441602 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-nova-metadata-neutron-config-0\") pod \"35d2beec-8391-420e-a413-44fed1095880\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.441785 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-metadata-combined-ca-bundle\") pod \"35d2beec-8391-420e-a413-44fed1095880\" (UID: \"35d2beec-8391-420e-a413-44fed1095880\") " Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.452081 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "35d2beec-8391-420e-a413-44fed1095880" (UID: "35d2beec-8391-420e-a413-44fed1095880"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.452194 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d2beec-8391-420e-a413-44fed1095880-kube-api-access-lpl4l" (OuterVolumeSpecName: "kube-api-access-lpl4l") pod "35d2beec-8391-420e-a413-44fed1095880" (UID: "35d2beec-8391-420e-a413-44fed1095880"). InnerVolumeSpecName "kube-api-access-lpl4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.473661 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "35d2beec-8391-420e-a413-44fed1095880" (UID: "35d2beec-8391-420e-a413-44fed1095880"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.473719 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35d2beec-8391-420e-a413-44fed1095880" (UID: "35d2beec-8391-420e-a413-44fed1095880"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.476845 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-inventory" (OuterVolumeSpecName: "inventory") pod "35d2beec-8391-420e-a413-44fed1095880" (UID: "35d2beec-8391-420e-a413-44fed1095880"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.487015 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "35d2beec-8391-420e-a413-44fed1095880" (UID: "35d2beec-8391-420e-a413-44fed1095880"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.544496 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.544535 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.544548 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpl4l\" (UniqueName: \"kubernetes.io/projected/35d2beec-8391-420e-a413-44fed1095880-kube-api-access-lpl4l\") on node \"crc\" DevicePath \"\"" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.544563 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.544574 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:20:00 crc kubenswrapper[4894]: I1208 15:20:00.544584 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/35d2beec-8391-420e-a413-44fed1095880-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.011142 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" event={"ID":"35d2beec-8391-420e-a413-44fed1095880","Type":"ContainerDied","Data":"f0db677c38ed7e23cd70d7b51df185d49e6d13d7267a9b2ed94a4eebd39ef442"} Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.011425 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0db677c38ed7e23cd70d7b51df185d49e6d13d7267a9b2ed94a4eebd39ef442" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.011197 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.172942 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7"] Dec 08 15:20:01 crc kubenswrapper[4894]: E1208 15:20:01.205113 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d2beec-8391-420e-a413-44fed1095880" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.205157 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d2beec-8391-420e-a413-44fed1095880" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.205906 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d2beec-8391-420e-a413-44fed1095880" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.207394 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.212253 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.212285 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.213019 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.213253 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.214616 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.238640 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7"] Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.255226 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.255486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzspf\" (UniqueName: \"kubernetes.io/projected/12612e53-bc51-4b09-9f07-f00d8debac18-kube-api-access-tzspf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.255551 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.255571 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.255587 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.357883 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.357950 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.357986 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.358133 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.358402 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzspf\" (UniqueName: \"kubernetes.io/projected/12612e53-bc51-4b09-9f07-f00d8debac18-kube-api-access-tzspf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.363220 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.363714 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.368500 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.371102 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.378483 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzspf\" (UniqueName: \"kubernetes.io/projected/12612e53-bc51-4b09-9f07-f00d8debac18-kube-api-access-tzspf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:01 crc kubenswrapper[4894]: I1208 15:20:01.535394 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:20:02 crc kubenswrapper[4894]: I1208 15:20:02.085528 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7"] Dec 08 15:20:03 crc kubenswrapper[4894]: I1208 15:20:03.033717 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" event={"ID":"12612e53-bc51-4b09-9f07-f00d8debac18","Type":"ContainerStarted","Data":"ca1ea0c57ba8d7a065b0904195f959cb43562e95b9fcd61cdc98bea87151ee8a"} Dec 08 15:20:03 crc kubenswrapper[4894]: I1208 15:20:03.034106 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" event={"ID":"12612e53-bc51-4b09-9f07-f00d8debac18","Type":"ContainerStarted","Data":"deee15891fcb5617968dddc6ddd0e5621b6b0af9a5f97500df3528b3693b403e"} Dec 08 15:20:03 crc kubenswrapper[4894]: I1208 15:20:03.054011 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" podStartSLOduration=1.592457824 podStartE2EDuration="2.053987701s" podCreationTimestamp="2025-12-08 15:20:01 +0000 UTC" firstStartedPulling="2025-12-08 15:20:02.089649913 +0000 UTC m=+2023.189656028" lastFinishedPulling="2025-12-08 15:20:02.55117978 +0000 UTC m=+2023.651185905" observedRunningTime="2025-12-08 15:20:03.047287871 +0000 UTC m=+2024.147294026" watchObservedRunningTime="2025-12-08 15:20:03.053987701 +0000 UTC m=+2024.153993836" Dec 08 15:20:07 crc kubenswrapper[4894]: I1208 15:20:07.297355 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:20:07 crc kubenswrapper[4894]: I1208 15:20:07.297775 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:20:37 crc kubenswrapper[4894]: I1208 15:20:37.297746 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:20:37 crc kubenswrapper[4894]: I1208 15:20:37.298473 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:20:37 crc kubenswrapper[4894]: I1208 15:20:37.298649 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:20:37 crc kubenswrapper[4894]: I1208 15:20:37.299668 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f23f6da90e9a9fd343cea89c0d4e56edbcbde0c11dbbfcf77ae9c0fffc82c8c3"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:20:37 crc kubenswrapper[4894]: I1208 15:20:37.299741 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://f23f6da90e9a9fd343cea89c0d4e56edbcbde0c11dbbfcf77ae9c0fffc82c8c3" gracePeriod=600 Dec 08 15:20:38 crc kubenswrapper[4894]: I1208 15:20:38.359746 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="f23f6da90e9a9fd343cea89c0d4e56edbcbde0c11dbbfcf77ae9c0fffc82c8c3" exitCode=0 Dec 08 15:20:38 crc kubenswrapper[4894]: I1208 15:20:38.359959 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"f23f6da90e9a9fd343cea89c0d4e56edbcbde0c11dbbfcf77ae9c0fffc82c8c3"} Dec 08 15:20:38 crc kubenswrapper[4894]: I1208 15:20:38.360336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0"} Dec 08 15:20:38 crc kubenswrapper[4894]: I1208 15:20:38.360374 4894 scope.go:117] "RemoveContainer" containerID="0c3505ed589ec715a88e08e283297f97a7db84bf68997b6ea3b509c8fca936fa" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.393453 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g7bzm"] Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.396236 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.405781 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g7bzm"] Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.524889 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b7zj\" (UniqueName: \"kubernetes.io/projected/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-kube-api-access-7b7zj\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.525366 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-utilities\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.525510 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-catalog-content\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.626927 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-utilities\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.627017 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-catalog-content\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.627050 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b7zj\" (UniqueName: \"kubernetes.io/projected/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-kube-api-access-7b7zj\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.627915 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-catalog-content\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.631160 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-utilities\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.671128 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b7zj\" (UniqueName: \"kubernetes.io/projected/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-kube-api-access-7b7zj\") pod \"community-operators-g7bzm\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:05 crc kubenswrapper[4894]: I1208 15:21:05.723826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:06 crc kubenswrapper[4894]: I1208 15:21:06.258557 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g7bzm"] Dec 08 15:21:06 crc kubenswrapper[4894]: I1208 15:21:06.641553 4894 generic.go:334] "Generic (PLEG): container finished" podID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerID="fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207" exitCode=0 Dec 08 15:21:06 crc kubenswrapper[4894]: I1208 15:21:06.641623 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7bzm" event={"ID":"dc0786d3-1b76-4036-a62b-a92e1bc31a3a","Type":"ContainerDied","Data":"fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207"} Dec 08 15:21:06 crc kubenswrapper[4894]: I1208 15:21:06.641835 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7bzm" event={"ID":"dc0786d3-1b76-4036-a62b-a92e1bc31a3a","Type":"ContainerStarted","Data":"abb94ad137036c9e61d5592296667f0c435d499322d00258ab442bd48d0a658d"} Dec 08 15:21:07 crc kubenswrapper[4894]: I1208 15:21:07.652240 4894 generic.go:334] "Generic (PLEG): container finished" podID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerID="a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2" exitCode=0 Dec 08 15:21:07 crc kubenswrapper[4894]: I1208 15:21:07.652293 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7bzm" event={"ID":"dc0786d3-1b76-4036-a62b-a92e1bc31a3a","Type":"ContainerDied","Data":"a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2"} Dec 08 15:21:08 crc kubenswrapper[4894]: I1208 15:21:08.664966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7bzm" event={"ID":"dc0786d3-1b76-4036-a62b-a92e1bc31a3a","Type":"ContainerStarted","Data":"4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb"} Dec 08 15:21:08 crc kubenswrapper[4894]: I1208 15:21:08.691013 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g7bzm" podStartSLOduration=2.3219136369999998 podStartE2EDuration="3.690988852s" podCreationTimestamp="2025-12-08 15:21:05 +0000 UTC" firstStartedPulling="2025-12-08 15:21:06.646910938 +0000 UTC m=+2087.746917063" lastFinishedPulling="2025-12-08 15:21:08.015986163 +0000 UTC m=+2089.115992278" observedRunningTime="2025-12-08 15:21:08.681021669 +0000 UTC m=+2089.781027784" watchObservedRunningTime="2025-12-08 15:21:08.690988852 +0000 UTC m=+2089.790994977" Dec 08 15:21:15 crc kubenswrapper[4894]: I1208 15:21:15.724096 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:15 crc kubenswrapper[4894]: I1208 15:21:15.725786 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:15 crc kubenswrapper[4894]: I1208 15:21:15.770630 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:16 crc kubenswrapper[4894]: I1208 15:21:16.773315 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:16 crc kubenswrapper[4894]: I1208 15:21:16.822899 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g7bzm"] Dec 08 15:21:18 crc kubenswrapper[4894]: I1208 15:21:18.753274 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g7bzm" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="registry-server" containerID="cri-o://4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb" gracePeriod=2 Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.215762 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.285125 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b7zj\" (UniqueName: \"kubernetes.io/projected/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-kube-api-access-7b7zj\") pod \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.285620 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-utilities\") pod \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.285664 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-catalog-content\") pod \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\" (UID: \"dc0786d3-1b76-4036-a62b-a92e1bc31a3a\") " Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.286434 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-utilities" (OuterVolumeSpecName: "utilities") pod "dc0786d3-1b76-4036-a62b-a92e1bc31a3a" (UID: "dc0786d3-1b76-4036-a62b-a92e1bc31a3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.290264 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-kube-api-access-7b7zj" (OuterVolumeSpecName: "kube-api-access-7b7zj") pod "dc0786d3-1b76-4036-a62b-a92e1bc31a3a" (UID: "dc0786d3-1b76-4036-a62b-a92e1bc31a3a"). InnerVolumeSpecName "kube-api-access-7b7zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.340406 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc0786d3-1b76-4036-a62b-a92e1bc31a3a" (UID: "dc0786d3-1b76-4036-a62b-a92e1bc31a3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.387020 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.387273 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b7zj\" (UniqueName: \"kubernetes.io/projected/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-kube-api-access-7b7zj\") on node \"crc\" DevicePath \"\"" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.387394 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0786d3-1b76-4036-a62b-a92e1bc31a3a-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.763801 4894 generic.go:334] "Generic (PLEG): container finished" podID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerID="4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb" exitCode=0 Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.763842 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7bzm" event={"ID":"dc0786d3-1b76-4036-a62b-a92e1bc31a3a","Type":"ContainerDied","Data":"4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb"} Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.763987 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g7bzm" event={"ID":"dc0786d3-1b76-4036-a62b-a92e1bc31a3a","Type":"ContainerDied","Data":"abb94ad137036c9e61d5592296667f0c435d499322d00258ab442bd48d0a658d"} Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.764014 4894 scope.go:117] "RemoveContainer" containerID="4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.764972 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g7bzm" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.818042 4894 scope.go:117] "RemoveContainer" containerID="a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.818521 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g7bzm"] Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.834726 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g7bzm"] Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.849465 4894 scope.go:117] "RemoveContainer" containerID="fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.889178 4894 scope.go:117] "RemoveContainer" containerID="4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb" Dec 08 15:21:19 crc kubenswrapper[4894]: E1208 15:21:19.889651 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb\": container with ID starting with 4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb not found: ID does not exist" containerID="4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.889691 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb"} err="failed to get container status \"4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb\": rpc error: code = NotFound desc = could not find container \"4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb\": container with ID starting with 4366b91eadee220cd615fa909a082114d99ac150dd60a6dc0649913167c2ffbb not found: ID does not exist" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.889716 4894 scope.go:117] "RemoveContainer" containerID="a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2" Dec 08 15:21:19 crc kubenswrapper[4894]: E1208 15:21:19.890151 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2\": container with ID starting with a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2 not found: ID does not exist" containerID="a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.890169 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2"} err="failed to get container status \"a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2\": rpc error: code = NotFound desc = could not find container \"a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2\": container with ID starting with a99e3dd93a301af54ba950018092e61c5aa6629c29b3da28996a35b0ad0ca5e2 not found: ID does not exist" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.890182 4894 scope.go:117] "RemoveContainer" containerID="fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207" Dec 08 15:21:19 crc kubenswrapper[4894]: E1208 15:21:19.890466 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207\": container with ID starting with fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207 not found: ID does not exist" containerID="fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207" Dec 08 15:21:19 crc kubenswrapper[4894]: I1208 15:21:19.890522 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207"} err="failed to get container status \"fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207\": rpc error: code = NotFound desc = could not find container \"fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207\": container with ID starting with fde392a50c975654974b339511bec68495cf14480c4894bca4f7a653aaff9207 not found: ID does not exist" Dec 08 15:21:21 crc kubenswrapper[4894]: I1208 15:21:21.208392 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" path="/var/lib/kubelet/pods/dc0786d3-1b76-4036-a62b-a92e1bc31a3a/volumes" Dec 08 15:22:37 crc kubenswrapper[4894]: I1208 15:22:37.297117 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:22:37 crc kubenswrapper[4894]: I1208 15:22:37.297591 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:23:07 crc kubenswrapper[4894]: I1208 15:23:07.297221 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:23:07 crc kubenswrapper[4894]: I1208 15:23:07.297741 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:23:37 crc kubenswrapper[4894]: I1208 15:23:37.296733 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:23:37 crc kubenswrapper[4894]: I1208 15:23:37.297328 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:23:37 crc kubenswrapper[4894]: I1208 15:23:37.297378 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:23:37 crc kubenswrapper[4894]: I1208 15:23:37.298146 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:23:37 crc kubenswrapper[4894]: I1208 15:23:37.298221 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" gracePeriod=600 Dec 08 15:23:37 crc kubenswrapper[4894]: E1208 15:23:37.429360 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:23:38 crc kubenswrapper[4894]: I1208 15:23:38.025339 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" exitCode=0 Dec 08 15:23:38 crc kubenswrapper[4894]: I1208 15:23:38.025523 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0"} Dec 08 15:23:38 crc kubenswrapper[4894]: I1208 15:23:38.025886 4894 scope.go:117] "RemoveContainer" containerID="f23f6da90e9a9fd343cea89c0d4e56edbcbde0c11dbbfcf77ae9c0fffc82c8c3" Dec 08 15:23:38 crc kubenswrapper[4894]: I1208 15:23:38.026777 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:23:38 crc kubenswrapper[4894]: E1208 15:23:38.027271 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:23:51 crc kubenswrapper[4894]: I1208 15:23:51.199638 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:23:51 crc kubenswrapper[4894]: E1208 15:23:51.201878 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:24:06 crc kubenswrapper[4894]: I1208 15:24:06.196524 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:24:06 crc kubenswrapper[4894]: E1208 15:24:06.197188 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:24:16 crc kubenswrapper[4894]: I1208 15:24:16.409150 4894 generic.go:334] "Generic (PLEG): container finished" podID="12612e53-bc51-4b09-9f07-f00d8debac18" containerID="ca1ea0c57ba8d7a065b0904195f959cb43562e95b9fcd61cdc98bea87151ee8a" exitCode=0 Dec 08 15:24:16 crc kubenswrapper[4894]: I1208 15:24:16.409258 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" event={"ID":"12612e53-bc51-4b09-9f07-f00d8debac18","Type":"ContainerDied","Data":"ca1ea0c57ba8d7a065b0904195f959cb43562e95b9fcd61cdc98bea87151ee8a"} Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.839941 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.987084 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-secret-0\") pod \"12612e53-bc51-4b09-9f07-f00d8debac18\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.987652 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-inventory\") pod \"12612e53-bc51-4b09-9f07-f00d8debac18\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.987767 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-ssh-key\") pod \"12612e53-bc51-4b09-9f07-f00d8debac18\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.988014 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzspf\" (UniqueName: \"kubernetes.io/projected/12612e53-bc51-4b09-9f07-f00d8debac18-kube-api-access-tzspf\") pod \"12612e53-bc51-4b09-9f07-f00d8debac18\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.988089 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-combined-ca-bundle\") pod \"12612e53-bc51-4b09-9f07-f00d8debac18\" (UID: \"12612e53-bc51-4b09-9f07-f00d8debac18\") " Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.994031 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12612e53-bc51-4b09-9f07-f00d8debac18-kube-api-access-tzspf" (OuterVolumeSpecName: "kube-api-access-tzspf") pod "12612e53-bc51-4b09-9f07-f00d8debac18" (UID: "12612e53-bc51-4b09-9f07-f00d8debac18"). InnerVolumeSpecName "kube-api-access-tzspf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:24:17 crc kubenswrapper[4894]: I1208 15:24:17.998296 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "12612e53-bc51-4b09-9f07-f00d8debac18" (UID: "12612e53-bc51-4b09-9f07-f00d8debac18"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.023180 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "12612e53-bc51-4b09-9f07-f00d8debac18" (UID: "12612e53-bc51-4b09-9f07-f00d8debac18"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.029041 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "12612e53-bc51-4b09-9f07-f00d8debac18" (UID: "12612e53-bc51-4b09-9f07-f00d8debac18"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.050631 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-inventory" (OuterVolumeSpecName: "inventory") pod "12612e53-bc51-4b09-9f07-f00d8debac18" (UID: "12612e53-bc51-4b09-9f07-f00d8debac18"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.091906 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzspf\" (UniqueName: \"kubernetes.io/projected/12612e53-bc51-4b09-9f07-f00d8debac18-kube-api-access-tzspf\") on node \"crc\" DevicePath \"\"" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.091958 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.091977 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.091992 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.092006 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12612e53-bc51-4b09-9f07-f00d8debac18-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.430798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" event={"ID":"12612e53-bc51-4b09-9f07-f00d8debac18","Type":"ContainerDied","Data":"deee15891fcb5617968dddc6ddd0e5621b6b0af9a5f97500df3528b3693b403e"} Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.430891 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="deee15891fcb5617968dddc6ddd0e5621b6b0af9a5f97500df3528b3693b403e" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.431117 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.514739 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz"] Dec 08 15:24:18 crc kubenswrapper[4894]: E1208 15:24:18.515120 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="extract-utilities" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.515134 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="extract-utilities" Dec 08 15:24:18 crc kubenswrapper[4894]: E1208 15:24:18.515143 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="registry-server" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.515151 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="registry-server" Dec 08 15:24:18 crc kubenswrapper[4894]: E1208 15:24:18.515159 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12612e53-bc51-4b09-9f07-f00d8debac18" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.515166 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="12612e53-bc51-4b09-9f07-f00d8debac18" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 08 15:24:18 crc kubenswrapper[4894]: E1208 15:24:18.515190 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="extract-content" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.515195 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="extract-content" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.515375 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="12612e53-bc51-4b09-9f07-f00d8debac18" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.515399 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc0786d3-1b76-4036-a62b-a92e1bc31a3a" containerName="registry-server" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.516007 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.518172 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.518288 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.518179 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.518915 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.527968 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz"] Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.557898 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.558365 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.558307 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601347 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601456 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601570 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601710 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p7fn\" (UniqueName: \"kubernetes.io/projected/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-kube-api-access-4p7fn\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601755 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601830 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.601927 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.602166 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.703924 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.704797 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.704985 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.705099 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.705283 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p7fn\" (UniqueName: \"kubernetes.io/projected/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-kube-api-access-4p7fn\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.705389 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.706315 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.706392 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.706672 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.706990 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.707990 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.708432 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.708709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.709447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.710227 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.710373 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.712990 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.720482 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p7fn\" (UniqueName: \"kubernetes.io/projected/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-kube-api-access-4p7fn\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ktkvz\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:18 crc kubenswrapper[4894]: I1208 15:24:18.874784 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:24:19 crc kubenswrapper[4894]: I1208 15:24:19.204150 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:24:19 crc kubenswrapper[4894]: E1208 15:24:19.204774 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:24:19 crc kubenswrapper[4894]: I1208 15:24:19.370868 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz"] Dec 08 15:24:19 crc kubenswrapper[4894]: I1208 15:24:19.378867 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:24:19 crc kubenswrapper[4894]: I1208 15:24:19.439653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" event={"ID":"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2","Type":"ContainerStarted","Data":"38899e6d2884c3f1764dbf817d8da64a252b6ee4fa2fd6188f676f2fd560c531"} Dec 08 15:24:19 crc kubenswrapper[4894]: I1208 15:24:19.858415 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:24:20 crc kubenswrapper[4894]: I1208 15:24:20.471373 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" event={"ID":"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2","Type":"ContainerStarted","Data":"2c7063ff4b7cb25b08b08a1cf549967682b99738fb83594402b30abdcf82c1e2"} Dec 08 15:24:20 crc kubenswrapper[4894]: I1208 15:24:20.495155 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" podStartSLOduration=2.018072394 podStartE2EDuration="2.495131885s" podCreationTimestamp="2025-12-08 15:24:18 +0000 UTC" firstStartedPulling="2025-12-08 15:24:19.378593185 +0000 UTC m=+2280.478599300" lastFinishedPulling="2025-12-08 15:24:19.855652676 +0000 UTC m=+2280.955658791" observedRunningTime="2025-12-08 15:24:20.493098742 +0000 UTC m=+2281.593104867" watchObservedRunningTime="2025-12-08 15:24:20.495131885 +0000 UTC m=+2281.595138020" Dec 08 15:24:34 crc kubenswrapper[4894]: I1208 15:24:34.197027 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:24:34 crc kubenswrapper[4894]: E1208 15:24:34.198447 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:24:49 crc kubenswrapper[4894]: I1208 15:24:49.209799 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:24:49 crc kubenswrapper[4894]: E1208 15:24:49.210498 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:25:02 crc kubenswrapper[4894]: I1208 15:25:02.197069 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:25:02 crc kubenswrapper[4894]: E1208 15:25:02.197744 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:25:17 crc kubenswrapper[4894]: I1208 15:25:17.198058 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:25:17 crc kubenswrapper[4894]: E1208 15:25:17.198952 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:25:28 crc kubenswrapper[4894]: I1208 15:25:28.197204 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:25:28 crc kubenswrapper[4894]: E1208 15:25:28.198145 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:25:39 crc kubenswrapper[4894]: I1208 15:25:39.202401 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:25:39 crc kubenswrapper[4894]: E1208 15:25:39.203115 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:25:54 crc kubenswrapper[4894]: I1208 15:25:54.198076 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:25:54 crc kubenswrapper[4894]: E1208 15:25:54.198755 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.717406 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9xgzw"] Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.720956 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.729610 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xgzw"] Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.817629 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmdbg\" (UniqueName: \"kubernetes.io/projected/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-kube-api-access-hmdbg\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.817953 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-utilities\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.818492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-catalog-content\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.919801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmdbg\" (UniqueName: \"kubernetes.io/projected/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-kube-api-access-hmdbg\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.920192 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-utilities\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.920441 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-catalog-content\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.920847 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-utilities\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.920887 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-catalog-content\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:56 crc kubenswrapper[4894]: I1208 15:25:56.943025 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmdbg\" (UniqueName: \"kubernetes.io/projected/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-kube-api-access-hmdbg\") pod \"redhat-marketplace-9xgzw\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:57 crc kubenswrapper[4894]: I1208 15:25:57.041239 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:25:57 crc kubenswrapper[4894]: I1208 15:25:57.590589 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xgzw"] Dec 08 15:25:58 crc kubenswrapper[4894]: I1208 15:25:58.408940 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerID="e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1" exitCode=0 Dec 08 15:25:58 crc kubenswrapper[4894]: I1208 15:25:58.408993 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xgzw" event={"ID":"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e","Type":"ContainerDied","Data":"e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1"} Dec 08 15:25:58 crc kubenswrapper[4894]: I1208 15:25:58.409225 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xgzw" event={"ID":"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e","Type":"ContainerStarted","Data":"a51d868e5c2d04b34da1d5202c0b7455c6996a1874f70b832b87d597582976e9"} Dec 08 15:25:59 crc kubenswrapper[4894]: I1208 15:25:59.419566 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerID="ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc" exitCode=0 Dec 08 15:25:59 crc kubenswrapper[4894]: I1208 15:25:59.419623 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xgzw" event={"ID":"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e","Type":"ContainerDied","Data":"ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc"} Dec 08 15:26:00 crc kubenswrapper[4894]: I1208 15:26:00.428852 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xgzw" event={"ID":"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e","Type":"ContainerStarted","Data":"51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7"} Dec 08 15:26:00 crc kubenswrapper[4894]: I1208 15:26:00.458598 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9xgzw" podStartSLOduration=3.03362129 podStartE2EDuration="4.45857784s" podCreationTimestamp="2025-12-08 15:25:56 +0000 UTC" firstStartedPulling="2025-12-08 15:25:58.411573498 +0000 UTC m=+2379.511579613" lastFinishedPulling="2025-12-08 15:25:59.836530048 +0000 UTC m=+2380.936536163" observedRunningTime="2025-12-08 15:26:00.452357355 +0000 UTC m=+2381.552363470" watchObservedRunningTime="2025-12-08 15:26:00.45857784 +0000 UTC m=+2381.558583955" Dec 08 15:26:07 crc kubenswrapper[4894]: I1208 15:26:07.041779 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:26:07 crc kubenswrapper[4894]: I1208 15:26:07.042313 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:26:07 crc kubenswrapper[4894]: I1208 15:26:07.093523 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:26:07 crc kubenswrapper[4894]: I1208 15:26:07.197536 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:26:07 crc kubenswrapper[4894]: E1208 15:26:07.197755 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:26:07 crc kubenswrapper[4894]: I1208 15:26:07.528675 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:26:07 crc kubenswrapper[4894]: I1208 15:26:07.585755 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xgzw"] Dec 08 15:26:09 crc kubenswrapper[4894]: I1208 15:26:09.506973 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9xgzw" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="registry-server" containerID="cri-o://51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7" gracePeriod=2 Dec 08 15:26:09 crc kubenswrapper[4894]: I1208 15:26:09.920514 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.073226 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-catalog-content\") pod \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.073353 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-utilities\") pod \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.073483 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmdbg\" (UniqueName: \"kubernetes.io/projected/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-kube-api-access-hmdbg\") pod \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\" (UID: \"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e\") " Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.074025 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-utilities" (OuterVolumeSpecName: "utilities") pod "2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" (UID: "2f9b8331-4ce0-4b35-883a-f5d8cf725f2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.078586 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-kube-api-access-hmdbg" (OuterVolumeSpecName: "kube-api-access-hmdbg") pod "2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" (UID: "2f9b8331-4ce0-4b35-883a-f5d8cf725f2e"). InnerVolumeSpecName "kube-api-access-hmdbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.092234 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" (UID: "2f9b8331-4ce0-4b35-883a-f5d8cf725f2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.175909 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmdbg\" (UniqueName: \"kubernetes.io/projected/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-kube-api-access-hmdbg\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.175935 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.175945 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.516027 4894 generic.go:334] "Generic (PLEG): container finished" podID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerID="51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7" exitCode=0 Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.516074 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xgzw" event={"ID":"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e","Type":"ContainerDied","Data":"51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7"} Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.516291 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xgzw" event={"ID":"2f9b8331-4ce0-4b35-883a-f5d8cf725f2e","Type":"ContainerDied","Data":"a51d868e5c2d04b34da1d5202c0b7455c6996a1874f70b832b87d597582976e9"} Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.516308 4894 scope.go:117] "RemoveContainer" containerID="51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.516091 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xgzw" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.549468 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xgzw"] Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.551710 4894 scope.go:117] "RemoveContainer" containerID="ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.560488 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xgzw"] Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.572299 4894 scope.go:117] "RemoveContainer" containerID="e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.614974 4894 scope.go:117] "RemoveContainer" containerID="51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7" Dec 08 15:26:10 crc kubenswrapper[4894]: E1208 15:26:10.615415 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7\": container with ID starting with 51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7 not found: ID does not exist" containerID="51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.615441 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7"} err="failed to get container status \"51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7\": rpc error: code = NotFound desc = could not find container \"51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7\": container with ID starting with 51e2fb4613786e457fd9fe61a801221a261c9d111736805b512ef2f1f9a34bc7 not found: ID does not exist" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.615462 4894 scope.go:117] "RemoveContainer" containerID="ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc" Dec 08 15:26:10 crc kubenswrapper[4894]: E1208 15:26:10.615841 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc\": container with ID starting with ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc not found: ID does not exist" containerID="ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.615888 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc"} err="failed to get container status \"ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc\": rpc error: code = NotFound desc = could not find container \"ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc\": container with ID starting with ab78ca53b6c4e798604465d74df97443733d5295a64d63ce0bf1d8171d10bcfc not found: ID does not exist" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.615913 4894 scope.go:117] "RemoveContainer" containerID="e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1" Dec 08 15:26:10 crc kubenswrapper[4894]: E1208 15:26:10.616181 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1\": container with ID starting with e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1 not found: ID does not exist" containerID="e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1" Dec 08 15:26:10 crc kubenswrapper[4894]: I1208 15:26:10.616204 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1"} err="failed to get container status \"e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1\": rpc error: code = NotFound desc = could not find container \"e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1\": container with ID starting with e87877c6ffe1aa5f9b733cb555664bbc8aa181d085ef83fd8d2e703dd3e870e1 not found: ID does not exist" Dec 08 15:26:11 crc kubenswrapper[4894]: I1208 15:26:11.206228 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" path="/var/lib/kubelet/pods/2f9b8331-4ce0-4b35-883a-f5d8cf725f2e/volumes" Dec 08 15:26:19 crc kubenswrapper[4894]: I1208 15:26:19.203673 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:26:19 crc kubenswrapper[4894]: E1208 15:26:19.204529 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.769307 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kn274"] Dec 08 15:26:23 crc kubenswrapper[4894]: E1208 15:26:23.770321 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="extract-utilities" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.770354 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="extract-utilities" Dec 08 15:26:23 crc kubenswrapper[4894]: E1208 15:26:23.770370 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="registry-server" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.770377 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="registry-server" Dec 08 15:26:23 crc kubenswrapper[4894]: E1208 15:26:23.770408 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="extract-content" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.770416 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="extract-content" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.770658 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f9b8331-4ce0-4b35-883a-f5d8cf725f2e" containerName="registry-server" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.772350 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.780596 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kn274"] Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.902185 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-utilities\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.902606 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-catalog-content\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:23 crc kubenswrapper[4894]: I1208 15:26:23.902659 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fndgl\" (UniqueName: \"kubernetes.io/projected/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-kube-api-access-fndgl\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.004144 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fndgl\" (UniqueName: \"kubernetes.io/projected/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-kube-api-access-fndgl\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.004193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-catalog-content\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.004287 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-utilities\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.004792 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-catalog-content\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.004857 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-utilities\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.022791 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fndgl\" (UniqueName: \"kubernetes.io/projected/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-kube-api-access-fndgl\") pod \"redhat-operators-kn274\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.100712 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.651242 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kn274"] Dec 08 15:26:24 crc kubenswrapper[4894]: I1208 15:26:24.681034 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn274" event={"ID":"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789","Type":"ContainerStarted","Data":"9afd630068fd7f7f180d64bf5b8064743563b5ba47fec41284b4afac9d900601"} Dec 08 15:26:25 crc kubenswrapper[4894]: I1208 15:26:25.690486 4894 generic.go:334] "Generic (PLEG): container finished" podID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerID="dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4" exitCode=0 Dec 08 15:26:25 crc kubenswrapper[4894]: I1208 15:26:25.690561 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn274" event={"ID":"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789","Type":"ContainerDied","Data":"dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4"} Dec 08 15:26:26 crc kubenswrapper[4894]: I1208 15:26:26.702420 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn274" event={"ID":"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789","Type":"ContainerStarted","Data":"8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725"} Dec 08 15:26:26 crc kubenswrapper[4894]: I1208 15:26:26.970717 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-82269"] Dec 08 15:26:26 crc kubenswrapper[4894]: I1208 15:26:26.976412 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:26 crc kubenswrapper[4894]: I1208 15:26:26.986699 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82269"] Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.164176 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g4gb\" (UniqueName: \"kubernetes.io/projected/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-kube-api-access-6g4gb\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.164248 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-utilities\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.164799 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-catalog-content\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.266348 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g4gb\" (UniqueName: \"kubernetes.io/projected/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-kube-api-access-6g4gb\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.266419 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-utilities\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.266497 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-catalog-content\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.267016 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-catalog-content\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.267287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-utilities\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.288627 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g4gb\" (UniqueName: \"kubernetes.io/projected/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-kube-api-access-6g4gb\") pod \"certified-operators-82269\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.303915 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:27 crc kubenswrapper[4894]: I1208 15:26:27.912666 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-82269"] Dec 08 15:26:28 crc kubenswrapper[4894]: I1208 15:26:28.722892 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82269" event={"ID":"53adbb76-6c33-40bb-9a0e-3f4ae976afcc","Type":"ContainerStarted","Data":"44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49"} Dec 08 15:26:28 crc kubenswrapper[4894]: I1208 15:26:28.723249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82269" event={"ID":"53adbb76-6c33-40bb-9a0e-3f4ae976afcc","Type":"ContainerStarted","Data":"ed3c1f15d7548142736aabc1dde3b73385e5fed86328ca6de367d948095f915a"} Dec 08 15:26:29 crc kubenswrapper[4894]: I1208 15:26:29.732070 4894 generic.go:334] "Generic (PLEG): container finished" podID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerID="44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49" exitCode=0 Dec 08 15:26:29 crc kubenswrapper[4894]: I1208 15:26:29.732176 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82269" event={"ID":"53adbb76-6c33-40bb-9a0e-3f4ae976afcc","Type":"ContainerDied","Data":"44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49"} Dec 08 15:26:29 crc kubenswrapper[4894]: I1208 15:26:29.735049 4894 generic.go:334] "Generic (PLEG): container finished" podID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerID="8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725" exitCode=0 Dec 08 15:26:29 crc kubenswrapper[4894]: I1208 15:26:29.735083 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn274" event={"ID":"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789","Type":"ContainerDied","Data":"8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725"} Dec 08 15:26:30 crc kubenswrapper[4894]: I1208 15:26:30.744009 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn274" event={"ID":"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789","Type":"ContainerStarted","Data":"26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c"} Dec 08 15:26:30 crc kubenswrapper[4894]: I1208 15:26:30.766746 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kn274" podStartSLOduration=3.307711446 podStartE2EDuration="7.766728338s" podCreationTimestamp="2025-12-08 15:26:23 +0000 UTC" firstStartedPulling="2025-12-08 15:26:25.692307788 +0000 UTC m=+2406.792313903" lastFinishedPulling="2025-12-08 15:26:30.15132468 +0000 UTC m=+2411.251330795" observedRunningTime="2025-12-08 15:26:30.76456595 +0000 UTC m=+2411.864572065" watchObservedRunningTime="2025-12-08 15:26:30.766728338 +0000 UTC m=+2411.866734453" Dec 08 15:26:31 crc kubenswrapper[4894]: I1208 15:26:31.752839 4894 generic.go:334] "Generic (PLEG): container finished" podID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerID="c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948" exitCode=0 Dec 08 15:26:31 crc kubenswrapper[4894]: I1208 15:26:31.752935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82269" event={"ID":"53adbb76-6c33-40bb-9a0e-3f4ae976afcc","Type":"ContainerDied","Data":"c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948"} Dec 08 15:26:32 crc kubenswrapper[4894]: I1208 15:26:32.198147 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:26:32 crc kubenswrapper[4894]: E1208 15:26:32.198938 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:26:32 crc kubenswrapper[4894]: I1208 15:26:32.761995 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82269" event={"ID":"53adbb76-6c33-40bb-9a0e-3f4ae976afcc","Type":"ContainerStarted","Data":"0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16"} Dec 08 15:26:32 crc kubenswrapper[4894]: I1208 15:26:32.792194 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-82269" podStartSLOduration=4.397228373 podStartE2EDuration="6.792174687s" podCreationTimestamp="2025-12-08 15:26:26 +0000 UTC" firstStartedPulling="2025-12-08 15:26:29.733726808 +0000 UTC m=+2410.833732923" lastFinishedPulling="2025-12-08 15:26:32.128673122 +0000 UTC m=+2413.228679237" observedRunningTime="2025-12-08 15:26:32.784500807 +0000 UTC m=+2413.884506922" watchObservedRunningTime="2025-12-08 15:26:32.792174687 +0000 UTC m=+2413.892180802" Dec 08 15:26:34 crc kubenswrapper[4894]: I1208 15:26:34.101386 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:34 crc kubenswrapper[4894]: I1208 15:26:34.101908 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:35 crc kubenswrapper[4894]: I1208 15:26:35.160672 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kn274" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="registry-server" probeResult="failure" output=< Dec 08 15:26:35 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 08 15:26:35 crc kubenswrapper[4894]: > Dec 08 15:26:37 crc kubenswrapper[4894]: I1208 15:26:37.304510 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:37 crc kubenswrapper[4894]: I1208 15:26:37.305102 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:37 crc kubenswrapper[4894]: I1208 15:26:37.353030 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:37 crc kubenswrapper[4894]: I1208 15:26:37.865522 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:37 crc kubenswrapper[4894]: I1208 15:26:37.960498 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82269"] Dec 08 15:26:39 crc kubenswrapper[4894]: I1208 15:26:39.823927 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-82269" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="registry-server" containerID="cri-o://0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16" gracePeriod=2 Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.296341 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.450761 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-utilities\") pod \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.451335 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g4gb\" (UniqueName: \"kubernetes.io/projected/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-kube-api-access-6g4gb\") pod \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.451954 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-utilities" (OuterVolumeSpecName: "utilities") pod "53adbb76-6c33-40bb-9a0e-3f4ae976afcc" (UID: "53adbb76-6c33-40bb-9a0e-3f4ae976afcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.452615 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-catalog-content\") pod \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\" (UID: \"53adbb76-6c33-40bb-9a0e-3f4ae976afcc\") " Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.453356 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.459555 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-kube-api-access-6g4gb" (OuterVolumeSpecName: "kube-api-access-6g4gb") pod "53adbb76-6c33-40bb-9a0e-3f4ae976afcc" (UID: "53adbb76-6c33-40bb-9a0e-3f4ae976afcc"). InnerVolumeSpecName "kube-api-access-6g4gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.512149 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53adbb76-6c33-40bb-9a0e-3f4ae976afcc" (UID: "53adbb76-6c33-40bb-9a0e-3f4ae976afcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.555120 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g4gb\" (UniqueName: \"kubernetes.io/projected/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-kube-api-access-6g4gb\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.555189 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53adbb76-6c33-40bb-9a0e-3f4ae976afcc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.849357 4894 generic.go:334] "Generic (PLEG): container finished" podID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerID="0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16" exitCode=0 Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.849414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82269" event={"ID":"53adbb76-6c33-40bb-9a0e-3f4ae976afcc","Type":"ContainerDied","Data":"0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16"} Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.849446 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-82269" event={"ID":"53adbb76-6c33-40bb-9a0e-3f4ae976afcc","Type":"ContainerDied","Data":"ed3c1f15d7548142736aabc1dde3b73385e5fed86328ca6de367d948095f915a"} Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.849445 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-82269" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.849460 4894 scope.go:117] "RemoveContainer" containerID="0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.872596 4894 scope.go:117] "RemoveContainer" containerID="c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.886111 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-82269"] Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.894017 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-82269"] Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.909063 4894 scope.go:117] "RemoveContainer" containerID="44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.954398 4894 scope.go:117] "RemoveContainer" containerID="0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16" Dec 08 15:26:40 crc kubenswrapper[4894]: E1208 15:26:40.954751 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16\": container with ID starting with 0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16 not found: ID does not exist" containerID="0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.954786 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16"} err="failed to get container status \"0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16\": rpc error: code = NotFound desc = could not find container \"0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16\": container with ID starting with 0fa7f93d79e80dabf3923b5ce6b98d816ad4532f89821d9938818fb2ec934d16 not found: ID does not exist" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.954808 4894 scope.go:117] "RemoveContainer" containerID="c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948" Dec 08 15:26:40 crc kubenswrapper[4894]: E1208 15:26:40.955032 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948\": container with ID starting with c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948 not found: ID does not exist" containerID="c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.955055 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948"} err="failed to get container status \"c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948\": rpc error: code = NotFound desc = could not find container \"c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948\": container with ID starting with c2d6e595486d79c0735bb803520521a260438aa413d24092b2b967e89851d948 not found: ID does not exist" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.955068 4894 scope.go:117] "RemoveContainer" containerID="44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49" Dec 08 15:26:40 crc kubenswrapper[4894]: E1208 15:26:40.955519 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49\": container with ID starting with 44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49 not found: ID does not exist" containerID="44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49" Dec 08 15:26:40 crc kubenswrapper[4894]: I1208 15:26:40.955547 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49"} err="failed to get container status \"44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49\": rpc error: code = NotFound desc = could not find container \"44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49\": container with ID starting with 44eca4593b0899da4586a4cc0ce93c82c22b907a6c512650e26353ef26e3ab49 not found: ID does not exist" Dec 08 15:26:41 crc kubenswrapper[4894]: I1208 15:26:41.208900 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" path="/var/lib/kubelet/pods/53adbb76-6c33-40bb-9a0e-3f4ae976afcc/volumes" Dec 08 15:26:43 crc kubenswrapper[4894]: I1208 15:26:43.196656 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:26:43 crc kubenswrapper[4894]: E1208 15:26:43.197309 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:26:44 crc kubenswrapper[4894]: I1208 15:26:44.145060 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:44 crc kubenswrapper[4894]: I1208 15:26:44.199413 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:44 crc kubenswrapper[4894]: I1208 15:26:44.381234 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kn274"] Dec 08 15:26:45 crc kubenswrapper[4894]: I1208 15:26:45.895447 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kn274" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="registry-server" containerID="cri-o://26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c" gracePeriod=2 Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.301350 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.468178 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fndgl\" (UniqueName: \"kubernetes.io/projected/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-kube-api-access-fndgl\") pod \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.468736 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-utilities\") pod \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.468879 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-catalog-content\") pod \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\" (UID: \"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789\") " Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.469167 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-utilities" (OuterVolumeSpecName: "utilities") pod "da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" (UID: "da3e7ae1-a02a-4610-8b1b-d38b8a4d5789"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.470280 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.474718 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-kube-api-access-fndgl" (OuterVolumeSpecName: "kube-api-access-fndgl") pod "da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" (UID: "da3e7ae1-a02a-4610-8b1b-d38b8a4d5789"). InnerVolumeSpecName "kube-api-access-fndgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.571929 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fndgl\" (UniqueName: \"kubernetes.io/projected/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-kube-api-access-fndgl\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.579510 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" (UID: "da3e7ae1-a02a-4610-8b1b-d38b8a4d5789"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.673600 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.908462 4894 generic.go:334] "Generic (PLEG): container finished" podID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerID="26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c" exitCode=0 Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.908519 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kn274" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.908520 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn274" event={"ID":"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789","Type":"ContainerDied","Data":"26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c"} Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.908705 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kn274" event={"ID":"da3e7ae1-a02a-4610-8b1b-d38b8a4d5789","Type":"ContainerDied","Data":"9afd630068fd7f7f180d64bf5b8064743563b5ba47fec41284b4afac9d900601"} Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.908744 4894 scope.go:117] "RemoveContainer" containerID="26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.947499 4894 scope.go:117] "RemoveContainer" containerID="8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725" Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.948693 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kn274"] Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.964259 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kn274"] Dec 08 15:26:46 crc kubenswrapper[4894]: I1208 15:26:46.989693 4894 scope.go:117] "RemoveContainer" containerID="dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4" Dec 08 15:26:47 crc kubenswrapper[4894]: I1208 15:26:47.034648 4894 scope.go:117] "RemoveContainer" containerID="26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c" Dec 08 15:26:47 crc kubenswrapper[4894]: E1208 15:26:47.035443 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c\": container with ID starting with 26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c not found: ID does not exist" containerID="26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c" Dec 08 15:26:47 crc kubenswrapper[4894]: I1208 15:26:47.035531 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c"} err="failed to get container status \"26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c\": rpc error: code = NotFound desc = could not find container \"26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c\": container with ID starting with 26ab4fdfdd45ae925de426316dbc3b1fd49567ce2f76738fa175ee4996b4bd4c not found: ID does not exist" Dec 08 15:26:47 crc kubenswrapper[4894]: I1208 15:26:47.035613 4894 scope.go:117] "RemoveContainer" containerID="8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725" Dec 08 15:26:47 crc kubenswrapper[4894]: E1208 15:26:47.036386 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725\": container with ID starting with 8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725 not found: ID does not exist" containerID="8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725" Dec 08 15:26:47 crc kubenswrapper[4894]: I1208 15:26:47.036456 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725"} err="failed to get container status \"8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725\": rpc error: code = NotFound desc = could not find container \"8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725\": container with ID starting with 8ffbc7dc699efd4ae0e9dc140ecdb0e67d1b0b822be5d5f21be211584ab8a725 not found: ID does not exist" Dec 08 15:26:47 crc kubenswrapper[4894]: I1208 15:26:47.036519 4894 scope.go:117] "RemoveContainer" containerID="dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4" Dec 08 15:26:47 crc kubenswrapper[4894]: E1208 15:26:47.036877 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4\": container with ID starting with dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4 not found: ID does not exist" containerID="dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4" Dec 08 15:26:47 crc kubenswrapper[4894]: I1208 15:26:47.036919 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4"} err="failed to get container status \"dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4\": rpc error: code = NotFound desc = could not find container \"dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4\": container with ID starting with dca927f13f5787ae4b4e6e5407a7d3a2abab0e5e74799d4e5d530216922259b4 not found: ID does not exist" Dec 08 15:26:47 crc kubenswrapper[4894]: I1208 15:26:47.207950 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" path="/var/lib/kubelet/pods/da3e7ae1-a02a-4610-8b1b-d38b8a4d5789/volumes" Dec 08 15:26:57 crc kubenswrapper[4894]: I1208 15:26:57.197728 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:26:57 crc kubenswrapper[4894]: E1208 15:26:57.198374 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:27:09 crc kubenswrapper[4894]: I1208 15:27:09.203336 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:27:09 crc kubenswrapper[4894]: E1208 15:27:09.204222 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:27:10 crc kubenswrapper[4894]: I1208 15:27:10.104863 4894 generic.go:334] "Generic (PLEG): container finished" podID="01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" containerID="2c7063ff4b7cb25b08b08a1cf549967682b99738fb83594402b30abdcf82c1e2" exitCode=0 Dec 08 15:27:10 crc kubenswrapper[4894]: I1208 15:27:10.104926 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" event={"ID":"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2","Type":"ContainerDied","Data":"2c7063ff4b7cb25b08b08a1cf549967682b99738fb83594402b30abdcf82c1e2"} Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.511874 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.627957 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-extra-config-0\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628007 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-1\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628039 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p7fn\" (UniqueName: \"kubernetes.io/projected/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-kube-api-access-4p7fn\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628121 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-0\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628171 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-ssh-key\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628217 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-1\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628244 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-0\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628313 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-inventory\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.628348 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-combined-ca-bundle\") pod \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\" (UID: \"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2\") " Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.635659 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-kube-api-access-4p7fn" (OuterVolumeSpecName: "kube-api-access-4p7fn") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "kube-api-access-4p7fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.657635 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.657660 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-inventory" (OuterVolumeSpecName: "inventory") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.662380 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.662807 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.667974 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.668021 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.678316 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.678798 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" (UID: "01439e20-f7a9-4aa6-bcd7-754e48a7d8b2"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.730909 4894 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731042 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731059 4894 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731071 4894 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731079 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731088 4894 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731096 4894 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731130 4894 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:11 crc kubenswrapper[4894]: I1208 15:27:11.731139 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p7fn\" (UniqueName: \"kubernetes.io/projected/01439e20-f7a9-4aa6-bcd7-754e48a7d8b2-kube-api-access-4p7fn\") on node \"crc\" DevicePath \"\"" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.120046 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" event={"ID":"01439e20-f7a9-4aa6-bcd7-754e48a7d8b2","Type":"ContainerDied","Data":"38899e6d2884c3f1764dbf817d8da64a252b6ee4fa2fd6188f676f2fd560c531"} Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.120306 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38899e6d2884c3f1764dbf817d8da64a252b6ee4fa2fd6188f676f2fd560c531" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.120106 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ktkvz" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215256 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp"] Dec 08 15:27:12 crc kubenswrapper[4894]: E1208 15:27:12.215713 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215737 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 08 15:27:12 crc kubenswrapper[4894]: E1208 15:27:12.215761 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="extract-content" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215771 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="extract-content" Dec 08 15:27:12 crc kubenswrapper[4894]: E1208 15:27:12.215802 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="registry-server" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215826 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="registry-server" Dec 08 15:27:12 crc kubenswrapper[4894]: E1208 15:27:12.215837 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="extract-content" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215844 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="extract-content" Dec 08 15:27:12 crc kubenswrapper[4894]: E1208 15:27:12.215859 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="extract-utilities" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215868 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="extract-utilities" Dec 08 15:27:12 crc kubenswrapper[4894]: E1208 15:27:12.215882 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="extract-utilities" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215889 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="extract-utilities" Dec 08 15:27:12 crc kubenswrapper[4894]: E1208 15:27:12.215907 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="registry-server" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.215916 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="registry-server" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.216199 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="53adbb76-6c33-40bb-9a0e-3f4ae976afcc" containerName="registry-server" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.216217 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="da3e7ae1-a02a-4610-8b1b-d38b8a4d5789" containerName="registry-server" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.216234 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="01439e20-f7a9-4aa6-bcd7-754e48a7d8b2" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.216958 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.221360 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.221531 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.221776 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.221960 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-s2m77" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.222137 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.226552 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp"] Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.343673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.343731 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.343857 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.343889 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.343962 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.344205 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.344397 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5nc4\" (UniqueName: \"kubernetes.io/projected/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-kube-api-access-t5nc4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.447053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.447142 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.447286 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.447358 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.447607 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5nc4\" (UniqueName: \"kubernetes.io/projected/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-kube-api-access-t5nc4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.447701 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.447762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.451968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.454685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.457923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.461177 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.462976 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.465325 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.489826 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5nc4\" (UniqueName: \"kubernetes.io/projected/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-kube-api-access-t5nc4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:12 crc kubenswrapper[4894]: I1208 15:27:12.534590 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:27:13 crc kubenswrapper[4894]: I1208 15:27:13.079770 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp"] Dec 08 15:27:13 crc kubenswrapper[4894]: I1208 15:27:13.129966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" event={"ID":"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6","Type":"ContainerStarted","Data":"0413557c547cf82603a4d1e96da7f57ebfd3f66d0a75611c638c9b7bc128a6c5"} Dec 08 15:27:14 crc kubenswrapper[4894]: I1208 15:27:14.140599 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" event={"ID":"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6","Type":"ContainerStarted","Data":"1c22bb4576c57469d83cc6d4755dd0600c888aba3099fe86afa34b71998e7d55"} Dec 08 15:27:14 crc kubenswrapper[4894]: I1208 15:27:14.160297 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" podStartSLOduration=1.7408678960000001 podStartE2EDuration="2.160275335s" podCreationTimestamp="2025-12-08 15:27:12 +0000 UTC" firstStartedPulling="2025-12-08 15:27:13.087304233 +0000 UTC m=+2454.187310348" lastFinishedPulling="2025-12-08 15:27:13.506711682 +0000 UTC m=+2454.606717787" observedRunningTime="2025-12-08 15:27:14.157836379 +0000 UTC m=+2455.257842494" watchObservedRunningTime="2025-12-08 15:27:14.160275335 +0000 UTC m=+2455.260281450" Dec 08 15:27:21 crc kubenswrapper[4894]: I1208 15:27:21.199708 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:27:21 crc kubenswrapper[4894]: E1208 15:27:21.203243 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:27:35 crc kubenswrapper[4894]: I1208 15:27:35.197318 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:27:35 crc kubenswrapper[4894]: E1208 15:27:35.198037 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:27:48 crc kubenswrapper[4894]: I1208 15:27:48.197180 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:27:48 crc kubenswrapper[4894]: E1208 15:27:48.199708 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:27:59 crc kubenswrapper[4894]: I1208 15:27:59.223917 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:27:59 crc kubenswrapper[4894]: E1208 15:27:59.225436 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:28:12 crc kubenswrapper[4894]: I1208 15:28:12.197438 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:28:12 crc kubenswrapper[4894]: E1208 15:28:12.198371 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:28:27 crc kubenswrapper[4894]: I1208 15:28:27.243494 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:28:27 crc kubenswrapper[4894]: E1208 15:28:27.244964 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:28:42 crc kubenswrapper[4894]: I1208 15:28:42.198349 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:28:42 crc kubenswrapper[4894]: I1208 15:28:42.949768 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"2b8ce01428bea0f208fe2d737243169003a793d035946f524114d8b7e719a094"} Dec 08 15:29:28 crc kubenswrapper[4894]: I1208 15:29:28.348923 4894 generic.go:334] "Generic (PLEG): container finished" podID="1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" containerID="1c22bb4576c57469d83cc6d4755dd0600c888aba3099fe86afa34b71998e7d55" exitCode=0 Dec 08 15:29:28 crc kubenswrapper[4894]: I1208 15:29:28.348983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" event={"ID":"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6","Type":"ContainerDied","Data":"1c22bb4576c57469d83cc6d4755dd0600c888aba3099fe86afa34b71998e7d55"} Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.784227 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.850727 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ssh-key\") pod \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.850779 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-1\") pod \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.850875 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-0\") pod \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.850947 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-2\") pod \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.851005 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-telemetry-combined-ca-bundle\") pod \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.851335 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5nc4\" (UniqueName: \"kubernetes.io/projected/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-kube-api-access-t5nc4\") pod \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.851418 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-inventory\") pod \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\" (UID: \"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6\") " Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.857386 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" (UID: "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.857428 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-kube-api-access-t5nc4" (OuterVolumeSpecName: "kube-api-access-t5nc4") pod "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" (UID: "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6"). InnerVolumeSpecName "kube-api-access-t5nc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.879928 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" (UID: "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.881486 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" (UID: "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.882276 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" (UID: "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.882421 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-inventory" (OuterVolumeSpecName: "inventory") pod "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" (UID: "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.883295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" (UID: "1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.953402 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5nc4\" (UniqueName: \"kubernetes.io/projected/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-kube-api-access-t5nc4\") on node \"crc\" DevicePath \"\"" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.953435 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-inventory\") on node \"crc\" DevicePath \"\"" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.953445 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.953455 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.953465 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.953474 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 08 15:29:29 crc kubenswrapper[4894]: I1208 15:29:29.953484 4894 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 08 15:29:30 crc kubenswrapper[4894]: I1208 15:29:30.377305 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" event={"ID":"1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6","Type":"ContainerDied","Data":"0413557c547cf82603a4d1e96da7f57ebfd3f66d0a75611c638c9b7bc128a6c5"} Dec 08 15:29:30 crc kubenswrapper[4894]: I1208 15:29:30.377357 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0413557c547cf82603a4d1e96da7f57ebfd3f66d0a75611c638c9b7bc128a6c5" Dec 08 15:29:30 crc kubenswrapper[4894]: I1208 15:29:30.377399 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.154414 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4"] Dec 08 15:30:00 crc kubenswrapper[4894]: E1208 15:30:00.156539 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.156671 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.157045 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.158076 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.163536 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.164065 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.169180 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4"] Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.333548 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6817badd-c26d-4cdc-8abf-e020e265678d-secret-volume\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.333740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6817badd-c26d-4cdc-8abf-e020e265678d-config-volume\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.333792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvkfw\" (UniqueName: \"kubernetes.io/projected/6817badd-c26d-4cdc-8abf-e020e265678d-kube-api-access-kvkfw\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.436302 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6817badd-c26d-4cdc-8abf-e020e265678d-config-volume\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.436492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvkfw\" (UniqueName: \"kubernetes.io/projected/6817badd-c26d-4cdc-8abf-e020e265678d-kube-api-access-kvkfw\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.436650 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6817badd-c26d-4cdc-8abf-e020e265678d-secret-volume\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.440758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6817badd-c26d-4cdc-8abf-e020e265678d-config-volume\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.448578 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6817badd-c26d-4cdc-8abf-e020e265678d-secret-volume\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.465589 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvkfw\" (UniqueName: \"kubernetes.io/projected/6817badd-c26d-4cdc-8abf-e020e265678d-kube-api-access-kvkfw\") pod \"collect-profiles-29420130-7tvp4\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.485551 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:00 crc kubenswrapper[4894]: I1208 15:30:00.941271 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4"] Dec 08 15:30:01 crc kubenswrapper[4894]: I1208 15:30:01.673493 4894 generic.go:334] "Generic (PLEG): container finished" podID="6817badd-c26d-4cdc-8abf-e020e265678d" containerID="5f52e6ab580a09f60106ecab25480b555fcd536e068e29d6c6a0c3cf1d628709" exitCode=0 Dec 08 15:30:01 crc kubenswrapper[4894]: I1208 15:30:01.673698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" event={"ID":"6817badd-c26d-4cdc-8abf-e020e265678d","Type":"ContainerDied","Data":"5f52e6ab580a09f60106ecab25480b555fcd536e068e29d6c6a0c3cf1d628709"} Dec 08 15:30:01 crc kubenswrapper[4894]: I1208 15:30:01.673861 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" event={"ID":"6817badd-c26d-4cdc-8abf-e020e265678d","Type":"ContainerStarted","Data":"0df8c854ff933ade50ea1970a36ce21cc5004e128c229d249c43d5ff8b7e8003"} Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.035857 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.217947 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6817badd-c26d-4cdc-8abf-e020e265678d-secret-volume\") pod \"6817badd-c26d-4cdc-8abf-e020e265678d\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.218065 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvkfw\" (UniqueName: \"kubernetes.io/projected/6817badd-c26d-4cdc-8abf-e020e265678d-kube-api-access-kvkfw\") pod \"6817badd-c26d-4cdc-8abf-e020e265678d\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.218086 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6817badd-c26d-4cdc-8abf-e020e265678d-config-volume\") pod \"6817badd-c26d-4cdc-8abf-e020e265678d\" (UID: \"6817badd-c26d-4cdc-8abf-e020e265678d\") " Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.219091 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6817badd-c26d-4cdc-8abf-e020e265678d-config-volume" (OuterVolumeSpecName: "config-volume") pod "6817badd-c26d-4cdc-8abf-e020e265678d" (UID: "6817badd-c26d-4cdc-8abf-e020e265678d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.224574 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6817badd-c26d-4cdc-8abf-e020e265678d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6817badd-c26d-4cdc-8abf-e020e265678d" (UID: "6817badd-c26d-4cdc-8abf-e020e265678d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.224926 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6817badd-c26d-4cdc-8abf-e020e265678d-kube-api-access-kvkfw" (OuterVolumeSpecName: "kube-api-access-kvkfw") pod "6817badd-c26d-4cdc-8abf-e020e265678d" (UID: "6817badd-c26d-4cdc-8abf-e020e265678d"). InnerVolumeSpecName "kube-api-access-kvkfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.320284 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvkfw\" (UniqueName: \"kubernetes.io/projected/6817badd-c26d-4cdc-8abf-e020e265678d-kube-api-access-kvkfw\") on node \"crc\" DevicePath \"\"" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.320318 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6817badd-c26d-4cdc-8abf-e020e265678d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.320327 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6817badd-c26d-4cdc-8abf-e020e265678d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.688903 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" event={"ID":"6817badd-c26d-4cdc-8abf-e020e265678d","Type":"ContainerDied","Data":"0df8c854ff933ade50ea1970a36ce21cc5004e128c229d249c43d5ff8b7e8003"} Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.688949 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0df8c854ff933ade50ea1970a36ce21cc5004e128c229d249c43d5ff8b7e8003" Dec 08 15:30:03 crc kubenswrapper[4894]: I1208 15:30:03.688953 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420130-7tvp4" Dec 08 15:30:04 crc kubenswrapper[4894]: I1208 15:30:04.122684 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth"] Dec 08 15:30:04 crc kubenswrapper[4894]: I1208 15:30:04.131208 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420085-47jth"] Dec 08 15:30:05 crc kubenswrapper[4894]: I1208 15:30:05.218586 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6dbcd2c-e683-4318-860b-447c2df78fdc" path="/var/lib/kubelet/pods/b6dbcd2c-e683-4318-860b-447c2df78fdc/volumes" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.810700 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 08 15:30:28 crc kubenswrapper[4894]: E1208 15:30:28.811738 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6817badd-c26d-4cdc-8abf-e020e265678d" containerName="collect-profiles" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.811752 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6817badd-c26d-4cdc-8abf-e020e265678d" containerName="collect-profiles" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.811983 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6817badd-c26d-4cdc-8abf-e020e265678d" containerName="collect-profiles" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.812732 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.815013 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.815286 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.815514 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-dq2lk" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.815705 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.875974 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.995410 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl6bp\" (UniqueName: \"kubernetes.io/projected/39de6187-9721-4687-8001-ef13c4abafb3-kube-api-access-kl6bp\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.995651 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.995675 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-config-data\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.995721 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.995735 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.995909 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.995953 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.996126 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:28 crc kubenswrapper[4894]: I1208 15:30:28.996194 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097499 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097565 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097615 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl6bp\" (UniqueName: \"kubernetes.io/projected/39de6187-9721-4687-8001-ef13c4abafb3-kube-api-access-kl6bp\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097666 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097686 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-config-data\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097727 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.097831 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.098027 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.098135 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.098278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.099058 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.099209 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-config-data\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.104389 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.104863 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.105370 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.115166 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl6bp\" (UniqueName: \"kubernetes.io/projected/39de6187-9721-4687-8001-ef13c4abafb3-kube-api-access-kl6bp\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.134545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.180032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.632292 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.640316 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:30:29 crc kubenswrapper[4894]: I1208 15:30:29.971470 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"39de6187-9721-4687-8001-ef13c4abafb3","Type":"ContainerStarted","Data":"974b3d602f8a19bc8dd6f4c96523d145d71f64fbab0b98a313a891796443b5a4"} Dec 08 15:30:58 crc kubenswrapper[4894]: I1208 15:30:58.084124 4894 scope.go:117] "RemoveContainer" containerID="e396f7b349155beb29b5750cb7eb4eebc6824895cab1545ce41f727124f1ea42" Dec 08 15:30:58 crc kubenswrapper[4894]: E1208 15:30:58.930948 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 08 15:30:58 crc kubenswrapper[4894]: E1208 15:30:58.931114 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kl6bp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(39de6187-9721-4687-8001-ef13c4abafb3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 08 15:30:58 crc kubenswrapper[4894]: E1208 15:30:58.932296 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="39de6187-9721-4687-8001-ef13c4abafb3" Dec 08 15:30:59 crc kubenswrapper[4894]: E1208 15:30:59.281029 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="39de6187-9721-4687-8001-ef13c4abafb3" Dec 08 15:31:07 crc kubenswrapper[4894]: I1208 15:31:07.296434 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:31:07 crc kubenswrapper[4894]: I1208 15:31:07.296750 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:31:16 crc kubenswrapper[4894]: I1208 15:31:16.450539 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"39de6187-9721-4687-8001-ef13c4abafb3","Type":"ContainerStarted","Data":"71f05a5c05e9306bfbe7afecb831d54408777c5adddd4df2b12f38584c049d2e"} Dec 08 15:31:16 crc kubenswrapper[4894]: I1208 15:31:16.486730 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.341288817 podStartE2EDuration="49.486704395s" podCreationTimestamp="2025-12-08 15:30:27 +0000 UTC" firstStartedPulling="2025-12-08 15:30:29.640023632 +0000 UTC m=+2650.740029757" lastFinishedPulling="2025-12-08 15:31:14.78543922 +0000 UTC m=+2695.885445335" observedRunningTime="2025-12-08 15:31:16.471166097 +0000 UTC m=+2697.571172222" watchObservedRunningTime="2025-12-08 15:31:16.486704395 +0000 UTC m=+2697.586710510" Dec 08 15:31:37 crc kubenswrapper[4894]: I1208 15:31:37.298098 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:31:37 crc kubenswrapper[4894]: I1208 15:31:37.298908 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:31:44 crc kubenswrapper[4894]: I1208 15:31:44.960557 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nbj8c"] Dec 08 15:31:44 crc kubenswrapper[4894]: I1208 15:31:44.962946 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:44 crc kubenswrapper[4894]: I1208 15:31:44.970489 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbj8c"] Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.086915 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-utilities\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.086961 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-catalog-content\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.087031 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6jxb\" (UniqueName: \"kubernetes.io/projected/febf650d-9a58-4381-b78b-bd79221d682b-kube-api-access-n6jxb\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.189214 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-utilities\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.189266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-catalog-content\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.189306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6jxb\" (UniqueName: \"kubernetes.io/projected/febf650d-9a58-4381-b78b-bd79221d682b-kube-api-access-n6jxb\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.189805 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-utilities\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.189979 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-catalog-content\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.227079 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6jxb\" (UniqueName: \"kubernetes.io/projected/febf650d-9a58-4381-b78b-bd79221d682b-kube-api-access-n6jxb\") pod \"community-operators-nbj8c\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.288285 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:45 crc kubenswrapper[4894]: I1208 15:31:45.880431 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbj8c"] Dec 08 15:31:46 crc kubenswrapper[4894]: I1208 15:31:46.819938 4894 generic.go:334] "Generic (PLEG): container finished" podID="febf650d-9a58-4381-b78b-bd79221d682b" containerID="30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57" exitCode=0 Dec 08 15:31:46 crc kubenswrapper[4894]: I1208 15:31:46.819985 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbj8c" event={"ID":"febf650d-9a58-4381-b78b-bd79221d682b","Type":"ContainerDied","Data":"30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57"} Dec 08 15:31:46 crc kubenswrapper[4894]: I1208 15:31:46.820017 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbj8c" event={"ID":"febf650d-9a58-4381-b78b-bd79221d682b","Type":"ContainerStarted","Data":"a0a76dd0ad40992391e152d716bf099c5ca13150614c5d446710c0f92bec3d50"} Dec 08 15:31:47 crc kubenswrapper[4894]: I1208 15:31:47.843498 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbj8c" event={"ID":"febf650d-9a58-4381-b78b-bd79221d682b","Type":"ContainerStarted","Data":"d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c"} Dec 08 15:31:48 crc kubenswrapper[4894]: I1208 15:31:48.852650 4894 generic.go:334] "Generic (PLEG): container finished" podID="febf650d-9a58-4381-b78b-bd79221d682b" containerID="d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c" exitCode=0 Dec 08 15:31:48 crc kubenswrapper[4894]: I1208 15:31:48.852698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbj8c" event={"ID":"febf650d-9a58-4381-b78b-bd79221d682b","Type":"ContainerDied","Data":"d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c"} Dec 08 15:31:49 crc kubenswrapper[4894]: I1208 15:31:49.866795 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbj8c" event={"ID":"febf650d-9a58-4381-b78b-bd79221d682b","Type":"ContainerStarted","Data":"924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7"} Dec 08 15:31:49 crc kubenswrapper[4894]: I1208 15:31:49.889544 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nbj8c" podStartSLOduration=3.475841553 podStartE2EDuration="5.889522384s" podCreationTimestamp="2025-12-08 15:31:44 +0000 UTC" firstStartedPulling="2025-12-08 15:31:46.821303799 +0000 UTC m=+2727.921309944" lastFinishedPulling="2025-12-08 15:31:49.23498466 +0000 UTC m=+2730.334990775" observedRunningTime="2025-12-08 15:31:49.881611966 +0000 UTC m=+2730.981618091" watchObservedRunningTime="2025-12-08 15:31:49.889522384 +0000 UTC m=+2730.989528499" Dec 08 15:31:55 crc kubenswrapper[4894]: I1208 15:31:55.288997 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:55 crc kubenswrapper[4894]: I1208 15:31:55.289521 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:55 crc kubenswrapper[4894]: I1208 15:31:55.344385 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:55 crc kubenswrapper[4894]: I1208 15:31:55.966407 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:56 crc kubenswrapper[4894]: I1208 15:31:56.014086 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbj8c"] Dec 08 15:31:57 crc kubenswrapper[4894]: I1208 15:31:57.949942 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nbj8c" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="registry-server" containerID="cri-o://924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7" gracePeriod=2 Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.470616 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.595169 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-utilities\") pod \"febf650d-9a58-4381-b78b-bd79221d682b\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.595235 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-catalog-content\") pod \"febf650d-9a58-4381-b78b-bd79221d682b\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.595403 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6jxb\" (UniqueName: \"kubernetes.io/projected/febf650d-9a58-4381-b78b-bd79221d682b-kube-api-access-n6jxb\") pod \"febf650d-9a58-4381-b78b-bd79221d682b\" (UID: \"febf650d-9a58-4381-b78b-bd79221d682b\") " Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.600155 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-utilities" (OuterVolumeSpecName: "utilities") pod "febf650d-9a58-4381-b78b-bd79221d682b" (UID: "febf650d-9a58-4381-b78b-bd79221d682b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.601299 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/febf650d-9a58-4381-b78b-bd79221d682b-kube-api-access-n6jxb" (OuterVolumeSpecName: "kube-api-access-n6jxb") pod "febf650d-9a58-4381-b78b-bd79221d682b" (UID: "febf650d-9a58-4381-b78b-bd79221d682b"). InnerVolumeSpecName "kube-api-access-n6jxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.644916 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "febf650d-9a58-4381-b78b-bd79221d682b" (UID: "febf650d-9a58-4381-b78b-bd79221d682b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.697559 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6jxb\" (UniqueName: \"kubernetes.io/projected/febf650d-9a58-4381-b78b-bd79221d682b-kube-api-access-n6jxb\") on node \"crc\" DevicePath \"\"" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.697592 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.697602 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febf650d-9a58-4381-b78b-bd79221d682b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.961232 4894 generic.go:334] "Generic (PLEG): container finished" podID="febf650d-9a58-4381-b78b-bd79221d682b" containerID="924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7" exitCode=0 Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.961289 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbj8c" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.961323 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbj8c" event={"ID":"febf650d-9a58-4381-b78b-bd79221d682b","Type":"ContainerDied","Data":"924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7"} Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.961639 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbj8c" event={"ID":"febf650d-9a58-4381-b78b-bd79221d682b","Type":"ContainerDied","Data":"a0a76dd0ad40992391e152d716bf099c5ca13150614c5d446710c0f92bec3d50"} Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.961663 4894 scope.go:117] "RemoveContainer" containerID="924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.988342 4894 scope.go:117] "RemoveContainer" containerID="d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c" Dec 08 15:31:58 crc kubenswrapper[4894]: I1208 15:31:58.995313 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbj8c"] Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.013345 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nbj8c"] Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.021844 4894 scope.go:117] "RemoveContainer" containerID="30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57" Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.061738 4894 scope.go:117] "RemoveContainer" containerID="924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7" Dec 08 15:31:59 crc kubenswrapper[4894]: E1208 15:31:59.062245 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7\": container with ID starting with 924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7 not found: ID does not exist" containerID="924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7" Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.062294 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7"} err="failed to get container status \"924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7\": rpc error: code = NotFound desc = could not find container \"924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7\": container with ID starting with 924ea14d80eaeb47e8a496270df4fc03955603c8d421f518c2de80dcb7497da7 not found: ID does not exist" Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.062325 4894 scope.go:117] "RemoveContainer" containerID="d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c" Dec 08 15:31:59 crc kubenswrapper[4894]: E1208 15:31:59.062716 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c\": container with ID starting with d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c not found: ID does not exist" containerID="d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c" Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.062750 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c"} err="failed to get container status \"d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c\": rpc error: code = NotFound desc = could not find container \"d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c\": container with ID starting with d8451f65ea600c927bfda6bd912d9fb362d10d66ce77c09f23cd1fbf4cd16b1c not found: ID does not exist" Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.062780 4894 scope.go:117] "RemoveContainer" containerID="30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57" Dec 08 15:31:59 crc kubenswrapper[4894]: E1208 15:31:59.063066 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57\": container with ID starting with 30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57 not found: ID does not exist" containerID="30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57" Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.063097 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57"} err="failed to get container status \"30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57\": rpc error: code = NotFound desc = could not find container \"30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57\": container with ID starting with 30d5daa814098832b726d2b7a098a42faedec48d2bc41f0daa8725f5e89bec57 not found: ID does not exist" Dec 08 15:31:59 crc kubenswrapper[4894]: I1208 15:31:59.207554 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="febf650d-9a58-4381-b78b-bd79221d682b" path="/var/lib/kubelet/pods/febf650d-9a58-4381-b78b-bd79221d682b/volumes" Dec 08 15:32:07 crc kubenswrapper[4894]: I1208 15:32:07.297070 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:32:07 crc kubenswrapper[4894]: I1208 15:32:07.297631 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:32:07 crc kubenswrapper[4894]: I1208 15:32:07.297698 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:32:07 crc kubenswrapper[4894]: I1208 15:32:07.298732 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b8ce01428bea0f208fe2d737243169003a793d035946f524114d8b7e719a094"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:32:07 crc kubenswrapper[4894]: I1208 15:32:07.298839 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://2b8ce01428bea0f208fe2d737243169003a793d035946f524114d8b7e719a094" gracePeriod=600 Dec 08 15:32:08 crc kubenswrapper[4894]: I1208 15:32:08.041050 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="2b8ce01428bea0f208fe2d737243169003a793d035946f524114d8b7e719a094" exitCode=0 Dec 08 15:32:08 crc kubenswrapper[4894]: I1208 15:32:08.041243 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"2b8ce01428bea0f208fe2d737243169003a793d035946f524114d8b7e719a094"} Dec 08 15:32:08 crc kubenswrapper[4894]: I1208 15:32:08.041621 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8"} Dec 08 15:32:08 crc kubenswrapper[4894]: I1208 15:32:08.041642 4894 scope.go:117] "RemoveContainer" containerID="c5f617e67ae65743db20b77e55767e23ece8a3d53c14bdb2ce29364d379c6bd0" Dec 08 15:34:07 crc kubenswrapper[4894]: I1208 15:34:07.296671 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:34:07 crc kubenswrapper[4894]: I1208 15:34:07.297221 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:34:37 crc kubenswrapper[4894]: I1208 15:34:37.296928 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:34:37 crc kubenswrapper[4894]: I1208 15:34:37.297407 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.297491 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.298147 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.298191 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.298717 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.298766 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" gracePeriod=600 Dec 08 15:35:07 crc kubenswrapper[4894]: E1208 15:35:07.419976 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.794259 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" exitCode=0 Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.794340 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8"} Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.794409 4894 scope.go:117] "RemoveContainer" containerID="2b8ce01428bea0f208fe2d737243169003a793d035946f524114d8b7e719a094" Dec 08 15:35:07 crc kubenswrapper[4894]: I1208 15:35:07.795702 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:35:07 crc kubenswrapper[4894]: E1208 15:35:07.796329 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:35:22 crc kubenswrapper[4894]: I1208 15:35:22.196872 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:35:22 crc kubenswrapper[4894]: E1208 15:35:22.197760 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:35:36 crc kubenswrapper[4894]: I1208 15:35:36.197010 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:35:36 crc kubenswrapper[4894]: E1208 15:35:36.197997 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:35:50 crc kubenswrapper[4894]: I1208 15:35:50.197252 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:35:50 crc kubenswrapper[4894]: E1208 15:35:50.198056 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.618233 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k94h2"] Dec 08 15:35:58 crc kubenswrapper[4894]: E1208 15:35:58.619269 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="registry-server" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.619291 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="registry-server" Dec 08 15:35:58 crc kubenswrapper[4894]: E1208 15:35:58.619331 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="extract-utilities" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.619340 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="extract-utilities" Dec 08 15:35:58 crc kubenswrapper[4894]: E1208 15:35:58.619362 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="extract-content" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.619370 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="extract-content" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.619623 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="febf650d-9a58-4381-b78b-bd79221d682b" containerName="registry-server" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.621617 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.629729 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94h2"] Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.680737 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-catalog-content\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.681246 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-utilities\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.681267 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvdq6\" (UniqueName: \"kubernetes.io/projected/210d0fc0-fc96-4700-a498-a50419bceb36-kube-api-access-nvdq6\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.783424 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-catalog-content\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.783503 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-utilities\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.783524 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvdq6\" (UniqueName: \"kubernetes.io/projected/210d0fc0-fc96-4700-a498-a50419bceb36-kube-api-access-nvdq6\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.784049 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-catalog-content\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.784092 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-utilities\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.806358 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvdq6\" (UniqueName: \"kubernetes.io/projected/210d0fc0-fc96-4700-a498-a50419bceb36-kube-api-access-nvdq6\") pod \"redhat-marketplace-k94h2\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:58 crc kubenswrapper[4894]: I1208 15:35:58.983165 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:35:59 crc kubenswrapper[4894]: I1208 15:35:59.445294 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94h2"] Dec 08 15:36:00 crc kubenswrapper[4894]: I1208 15:36:00.349167 4894 generic.go:334] "Generic (PLEG): container finished" podID="210d0fc0-fc96-4700-a498-a50419bceb36" containerID="1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060" exitCode=0 Dec 08 15:36:00 crc kubenswrapper[4894]: I1208 15:36:00.349219 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94h2" event={"ID":"210d0fc0-fc96-4700-a498-a50419bceb36","Type":"ContainerDied","Data":"1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060"} Dec 08 15:36:00 crc kubenswrapper[4894]: I1208 15:36:00.349490 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94h2" event={"ID":"210d0fc0-fc96-4700-a498-a50419bceb36","Type":"ContainerStarted","Data":"7eb14852bb584feab133925dc806ff8bca17835dfccf19a8ea44f0f1dc97315c"} Dec 08 15:36:00 crc kubenswrapper[4894]: I1208 15:36:00.351253 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:36:01 crc kubenswrapper[4894]: I1208 15:36:01.358873 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94h2" event={"ID":"210d0fc0-fc96-4700-a498-a50419bceb36","Type":"ContainerStarted","Data":"ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c"} Dec 08 15:36:02 crc kubenswrapper[4894]: I1208 15:36:02.368910 4894 generic.go:334] "Generic (PLEG): container finished" podID="210d0fc0-fc96-4700-a498-a50419bceb36" containerID="ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c" exitCode=0 Dec 08 15:36:02 crc kubenswrapper[4894]: I1208 15:36:02.369217 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94h2" event={"ID":"210d0fc0-fc96-4700-a498-a50419bceb36","Type":"ContainerDied","Data":"ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c"} Dec 08 15:36:03 crc kubenswrapper[4894]: I1208 15:36:03.197665 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:36:03 crc kubenswrapper[4894]: E1208 15:36:03.198232 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:36:03 crc kubenswrapper[4894]: I1208 15:36:03.385774 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94h2" event={"ID":"210d0fc0-fc96-4700-a498-a50419bceb36","Type":"ContainerStarted","Data":"90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab"} Dec 08 15:36:03 crc kubenswrapper[4894]: I1208 15:36:03.418081 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k94h2" podStartSLOduration=3.010762311 podStartE2EDuration="5.418058693s" podCreationTimestamp="2025-12-08 15:35:58 +0000 UTC" firstStartedPulling="2025-12-08 15:36:00.351025992 +0000 UTC m=+2981.451032107" lastFinishedPulling="2025-12-08 15:36:02.758322354 +0000 UTC m=+2983.858328489" observedRunningTime="2025-12-08 15:36:03.40965583 +0000 UTC m=+2984.509661945" watchObservedRunningTime="2025-12-08 15:36:03.418058693 +0000 UTC m=+2984.518064798" Dec 08 15:36:08 crc kubenswrapper[4894]: I1208 15:36:08.984001 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:36:08 crc kubenswrapper[4894]: I1208 15:36:08.984540 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:36:09 crc kubenswrapper[4894]: I1208 15:36:09.033333 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:36:09 crc kubenswrapper[4894]: I1208 15:36:09.532774 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:36:10 crc kubenswrapper[4894]: I1208 15:36:10.275417 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94h2"] Dec 08 15:36:11 crc kubenswrapper[4894]: I1208 15:36:11.490184 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k94h2" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="registry-server" containerID="cri-o://90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab" gracePeriod=2 Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.023588 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.143438 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-utilities\") pod \"210d0fc0-fc96-4700-a498-a50419bceb36\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.143504 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-catalog-content\") pod \"210d0fc0-fc96-4700-a498-a50419bceb36\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.143580 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvdq6\" (UniqueName: \"kubernetes.io/projected/210d0fc0-fc96-4700-a498-a50419bceb36-kube-api-access-nvdq6\") pod \"210d0fc0-fc96-4700-a498-a50419bceb36\" (UID: \"210d0fc0-fc96-4700-a498-a50419bceb36\") " Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.145020 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-utilities" (OuterVolumeSpecName: "utilities") pod "210d0fc0-fc96-4700-a498-a50419bceb36" (UID: "210d0fc0-fc96-4700-a498-a50419bceb36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.150081 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d0fc0-fc96-4700-a498-a50419bceb36-kube-api-access-nvdq6" (OuterVolumeSpecName: "kube-api-access-nvdq6") pod "210d0fc0-fc96-4700-a498-a50419bceb36" (UID: "210d0fc0-fc96-4700-a498-a50419bceb36"). InnerVolumeSpecName "kube-api-access-nvdq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.162934 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "210d0fc0-fc96-4700-a498-a50419bceb36" (UID: "210d0fc0-fc96-4700-a498-a50419bceb36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.246604 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvdq6\" (UniqueName: \"kubernetes.io/projected/210d0fc0-fc96-4700-a498-a50419bceb36-kube-api-access-nvdq6\") on node \"crc\" DevicePath \"\"" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.246639 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.246653 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/210d0fc0-fc96-4700-a498-a50419bceb36-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.498654 4894 generic.go:334] "Generic (PLEG): container finished" podID="210d0fc0-fc96-4700-a498-a50419bceb36" containerID="90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab" exitCode=0 Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.498706 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k94h2" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.498704 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94h2" event={"ID":"210d0fc0-fc96-4700-a498-a50419bceb36","Type":"ContainerDied","Data":"90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab"} Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.498871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k94h2" event={"ID":"210d0fc0-fc96-4700-a498-a50419bceb36","Type":"ContainerDied","Data":"7eb14852bb584feab133925dc806ff8bca17835dfccf19a8ea44f0f1dc97315c"} Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.498904 4894 scope.go:117] "RemoveContainer" containerID="90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.520068 4894 scope.go:117] "RemoveContainer" containerID="ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.532157 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94h2"] Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.544175 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k94h2"] Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.559671 4894 scope.go:117] "RemoveContainer" containerID="1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.584134 4894 scope.go:117] "RemoveContainer" containerID="90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab" Dec 08 15:36:12 crc kubenswrapper[4894]: E1208 15:36:12.584852 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab\": container with ID starting with 90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab not found: ID does not exist" containerID="90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.584936 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab"} err="failed to get container status \"90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab\": rpc error: code = NotFound desc = could not find container \"90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab\": container with ID starting with 90fe0763c84889da345a4daff7eb5c317636607d52a7e784970e95dd1bc987ab not found: ID does not exist" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.585009 4894 scope.go:117] "RemoveContainer" containerID="ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c" Dec 08 15:36:12 crc kubenswrapper[4894]: E1208 15:36:12.585350 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c\": container with ID starting with ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c not found: ID does not exist" containerID="ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.585380 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c"} err="failed to get container status \"ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c\": rpc error: code = NotFound desc = could not find container \"ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c\": container with ID starting with ff73b07706234e5577486cddbca21b1952416a4a2d88d25e15d9c9b40e181b2c not found: ID does not exist" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.585403 4894 scope.go:117] "RemoveContainer" containerID="1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060" Dec 08 15:36:12 crc kubenswrapper[4894]: E1208 15:36:12.585660 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060\": container with ID starting with 1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060 not found: ID does not exist" containerID="1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060" Dec 08 15:36:12 crc kubenswrapper[4894]: I1208 15:36:12.585686 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060"} err="failed to get container status \"1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060\": rpc error: code = NotFound desc = could not find container \"1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060\": container with ID starting with 1d9622987391ab8eb3fb9af7dc0c338ce71b4a06087320084062814938075060 not found: ID does not exist" Dec 08 15:36:13 crc kubenswrapper[4894]: I1208 15:36:13.211511 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" path="/var/lib/kubelet/pods/210d0fc0-fc96-4700-a498-a50419bceb36/volumes" Dec 08 15:36:16 crc kubenswrapper[4894]: I1208 15:36:16.196491 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:36:16 crc kubenswrapper[4894]: E1208 15:36:16.197091 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:36:29 crc kubenswrapper[4894]: I1208 15:36:29.208948 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:36:29 crc kubenswrapper[4894]: E1208 15:36:29.210321 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.013708 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rzwn9"] Dec 08 15:36:32 crc kubenswrapper[4894]: E1208 15:36:32.014754 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="registry-server" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.014770 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="registry-server" Dec 08 15:36:32 crc kubenswrapper[4894]: E1208 15:36:32.014791 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="extract-content" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.014797 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="extract-content" Dec 08 15:36:32 crc kubenswrapper[4894]: E1208 15:36:32.014838 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="extract-utilities" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.014845 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="extract-utilities" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.015043 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="210d0fc0-fc96-4700-a498-a50419bceb36" containerName="registry-server" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.016414 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.025408 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzwn9"] Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.173703 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k96d5\" (UniqueName: \"kubernetes.io/projected/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-kube-api-access-k96d5\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.173906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-catalog-content\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.173973 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-utilities\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.275787 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-catalog-content\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.276432 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-catalog-content\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.276592 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-utilities\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.277067 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-utilities\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.277211 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k96d5\" (UniqueName: \"kubernetes.io/projected/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-kube-api-access-k96d5\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.300270 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k96d5\" (UniqueName: \"kubernetes.io/projected/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-kube-api-access-k96d5\") pod \"redhat-operators-rzwn9\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.350389 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:32 crc kubenswrapper[4894]: I1208 15:36:32.823016 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzwn9"] Dec 08 15:36:32 crc kubenswrapper[4894]: W1208 15:36:32.835003 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0bc4ebb_0fde_40d1_8ddd_4e83fd85fa6c.slice/crio-21ad4049344dd5ab87e6b956e64126f62c065e3b1ab4391d81c3da1b859b522b WatchSource:0}: Error finding container 21ad4049344dd5ab87e6b956e64126f62c065e3b1ab4391d81c3da1b859b522b: Status 404 returned error can't find the container with id 21ad4049344dd5ab87e6b956e64126f62c065e3b1ab4391d81c3da1b859b522b Dec 08 15:36:33 crc kubenswrapper[4894]: I1208 15:36:33.780402 4894 generic.go:334] "Generic (PLEG): container finished" podID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerID="1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430" exitCode=0 Dec 08 15:36:33 crc kubenswrapper[4894]: I1208 15:36:33.780498 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzwn9" event={"ID":"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c","Type":"ContainerDied","Data":"1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430"} Dec 08 15:36:33 crc kubenswrapper[4894]: I1208 15:36:33.780984 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzwn9" event={"ID":"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c","Type":"ContainerStarted","Data":"21ad4049344dd5ab87e6b956e64126f62c065e3b1ab4391d81c3da1b859b522b"} Dec 08 15:36:34 crc kubenswrapper[4894]: I1208 15:36:34.791447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzwn9" event={"ID":"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c","Type":"ContainerStarted","Data":"3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c"} Dec 08 15:36:36 crc kubenswrapper[4894]: I1208 15:36:36.814061 4894 generic.go:334] "Generic (PLEG): container finished" podID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerID="3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c" exitCode=0 Dec 08 15:36:36 crc kubenswrapper[4894]: I1208 15:36:36.814158 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzwn9" event={"ID":"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c","Type":"ContainerDied","Data":"3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c"} Dec 08 15:36:37 crc kubenswrapper[4894]: I1208 15:36:37.868279 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzwn9" event={"ID":"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c","Type":"ContainerStarted","Data":"25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12"} Dec 08 15:36:37 crc kubenswrapper[4894]: I1208 15:36:37.896092 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rzwn9" podStartSLOduration=3.137607035 podStartE2EDuration="6.896070205s" podCreationTimestamp="2025-12-08 15:36:31 +0000 UTC" firstStartedPulling="2025-12-08 15:36:33.783536854 +0000 UTC m=+3014.883542999" lastFinishedPulling="2025-12-08 15:36:37.542000014 +0000 UTC m=+3018.642006169" observedRunningTime="2025-12-08 15:36:37.887874379 +0000 UTC m=+3018.987880494" watchObservedRunningTime="2025-12-08 15:36:37.896070205 +0000 UTC m=+3018.996076330" Dec 08 15:36:42 crc kubenswrapper[4894]: I1208 15:36:42.351668 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:42 crc kubenswrapper[4894]: I1208 15:36:42.352198 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:43 crc kubenswrapper[4894]: I1208 15:36:43.398034 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rzwn9" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="registry-server" probeResult="failure" output=< Dec 08 15:36:43 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 08 15:36:43 crc kubenswrapper[4894]: > Dec 08 15:36:44 crc kubenswrapper[4894]: I1208 15:36:44.196251 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:36:44 crc kubenswrapper[4894]: E1208 15:36:44.196493 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:36:52 crc kubenswrapper[4894]: I1208 15:36:52.411421 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:52 crc kubenswrapper[4894]: I1208 15:36:52.470298 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:52 crc kubenswrapper[4894]: I1208 15:36:52.652896 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rzwn9"] Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.013321 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rzwn9" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="registry-server" containerID="cri-o://25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12" gracePeriod=2 Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.524192 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.639042 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k96d5\" (UniqueName: \"kubernetes.io/projected/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-kube-api-access-k96d5\") pod \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.639112 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-utilities\") pod \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.639254 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-catalog-content\") pod \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\" (UID: \"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c\") " Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.640528 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-utilities" (OuterVolumeSpecName: "utilities") pod "d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" (UID: "d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.645417 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-kube-api-access-k96d5" (OuterVolumeSpecName: "kube-api-access-k96d5") pod "d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" (UID: "d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c"). InnerVolumeSpecName "kube-api-access-k96d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.739735 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" (UID: "d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.741732 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.741767 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k96d5\" (UniqueName: \"kubernetes.io/projected/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-kube-api-access-k96d5\") on node \"crc\" DevicePath \"\"" Dec 08 15:36:54 crc kubenswrapper[4894]: I1208 15:36:54.741779 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.024431 4894 generic.go:334] "Generic (PLEG): container finished" podID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerID="25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12" exitCode=0 Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.024469 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzwn9" event={"ID":"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c","Type":"ContainerDied","Data":"25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12"} Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.024496 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzwn9" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.024518 4894 scope.go:117] "RemoveContainer" containerID="25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.024506 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzwn9" event={"ID":"d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c","Type":"ContainerDied","Data":"21ad4049344dd5ab87e6b956e64126f62c065e3b1ab4391d81c3da1b859b522b"} Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.060310 4894 scope.go:117] "RemoveContainer" containerID="3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.071546 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rzwn9"] Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.079377 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rzwn9"] Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.082224 4894 scope.go:117] "RemoveContainer" containerID="1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.122939 4894 scope.go:117] "RemoveContainer" containerID="25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12" Dec 08 15:36:55 crc kubenswrapper[4894]: E1208 15:36:55.123618 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12\": container with ID starting with 25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12 not found: ID does not exist" containerID="25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.123657 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12"} err="failed to get container status \"25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12\": rpc error: code = NotFound desc = could not find container \"25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12\": container with ID starting with 25e29edfc99ef1ceea6a7028e592623828f54d709fed6494dfa03e0e616d5e12 not found: ID does not exist" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.123683 4894 scope.go:117] "RemoveContainer" containerID="3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c" Dec 08 15:36:55 crc kubenswrapper[4894]: E1208 15:36:55.123990 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c\": container with ID starting with 3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c not found: ID does not exist" containerID="3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.124015 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c"} err="failed to get container status \"3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c\": rpc error: code = NotFound desc = could not find container \"3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c\": container with ID starting with 3c71ed7ac70c6ba0b5a64dcb6e994afa90b7679e64f0fd568e4f4c7953f0406c not found: ID does not exist" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.124034 4894 scope.go:117] "RemoveContainer" containerID="1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430" Dec 08 15:36:55 crc kubenswrapper[4894]: E1208 15:36:55.124393 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430\": container with ID starting with 1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430 not found: ID does not exist" containerID="1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.124501 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430"} err="failed to get container status \"1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430\": rpc error: code = NotFound desc = could not find container \"1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430\": container with ID starting with 1089344e3ae93a54ce4a78ffd664d1b176d9b9be6c89461f56ea022e4d71f430 not found: ID does not exist" Dec 08 15:36:55 crc kubenswrapper[4894]: I1208 15:36:55.209589 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" path="/var/lib/kubelet/pods/d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c/volumes" Dec 08 15:36:57 crc kubenswrapper[4894]: I1208 15:36:57.197179 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:36:57 crc kubenswrapper[4894]: E1208 15:36:57.197700 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:37:09 crc kubenswrapper[4894]: I1208 15:37:09.223408 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:37:09 crc kubenswrapper[4894]: E1208 15:37:09.224395 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:37:22 crc kubenswrapper[4894]: I1208 15:37:22.196342 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:37:22 crc kubenswrapper[4894]: E1208 15:37:22.197126 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:37:35 crc kubenswrapper[4894]: I1208 15:37:35.196842 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:37:35 crc kubenswrapper[4894]: E1208 15:37:35.197552 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:37:46 crc kubenswrapper[4894]: I1208 15:37:46.196837 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:37:46 crc kubenswrapper[4894]: E1208 15:37:46.197541 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:37:58 crc kubenswrapper[4894]: I1208 15:37:58.196639 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:37:58 crc kubenswrapper[4894]: E1208 15:37:58.197332 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:38:10 crc kubenswrapper[4894]: I1208 15:38:10.197405 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:38:10 crc kubenswrapper[4894]: E1208 15:38:10.198698 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:38:21 crc kubenswrapper[4894]: I1208 15:38:21.197016 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:38:21 crc kubenswrapper[4894]: E1208 15:38:21.197800 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:38:33 crc kubenswrapper[4894]: I1208 15:38:33.197721 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:38:33 crc kubenswrapper[4894]: E1208 15:38:33.198960 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:38:48 crc kubenswrapper[4894]: I1208 15:38:48.198535 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:38:48 crc kubenswrapper[4894]: E1208 15:38:48.199776 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:39:00 crc kubenswrapper[4894]: I1208 15:39:00.199317 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:39:00 crc kubenswrapper[4894]: E1208 15:39:00.200424 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:39:14 crc kubenswrapper[4894]: I1208 15:39:14.197527 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:39:14 crc kubenswrapper[4894]: E1208 15:39:14.198251 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:39:28 crc kubenswrapper[4894]: I1208 15:39:28.197153 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:39:28 crc kubenswrapper[4894]: E1208 15:39:28.198182 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:39:39 crc kubenswrapper[4894]: I1208 15:39:39.203808 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:39:39 crc kubenswrapper[4894]: E1208 15:39:39.204685 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.897254 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c4z2f"] Dec 08 15:39:47 crc kubenswrapper[4894]: E1208 15:39:47.898207 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="extract-content" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.898221 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="extract-content" Dec 08 15:39:47 crc kubenswrapper[4894]: E1208 15:39:47.898242 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="registry-server" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.898248 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="registry-server" Dec 08 15:39:47 crc kubenswrapper[4894]: E1208 15:39:47.898259 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="extract-utilities" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.898265 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="extract-utilities" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.898436 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0bc4ebb-0fde-40d1-8ddd-4e83fd85fa6c" containerName="registry-server" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.899938 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.919165 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4z2f"] Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.946326 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-catalog-content\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.946639 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-utilities\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:47 crc kubenswrapper[4894]: I1208 15:39:47.946793 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szrxr\" (UniqueName: \"kubernetes.io/projected/e5806d93-34c4-41ab-944f-9e9507afe1a0-kube-api-access-szrxr\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.047755 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-utilities\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.047909 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szrxr\" (UniqueName: \"kubernetes.io/projected/e5806d93-34c4-41ab-944f-9e9507afe1a0-kube-api-access-szrxr\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.047998 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-catalog-content\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.048761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-catalog-content\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.049119 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-utilities\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.085250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szrxr\" (UniqueName: \"kubernetes.io/projected/e5806d93-34c4-41ab-944f-9e9507afe1a0-kube-api-access-szrxr\") pod \"certified-operators-c4z2f\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.248880 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:48 crc kubenswrapper[4894]: I1208 15:39:48.789267 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4z2f"] Dec 08 15:39:49 crc kubenswrapper[4894]: I1208 15:39:49.624338 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerID="105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d" exitCode=0 Dec 08 15:39:49 crc kubenswrapper[4894]: I1208 15:39:49.624410 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4z2f" event={"ID":"e5806d93-34c4-41ab-944f-9e9507afe1a0","Type":"ContainerDied","Data":"105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d"} Dec 08 15:39:49 crc kubenswrapper[4894]: I1208 15:39:49.624710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4z2f" event={"ID":"e5806d93-34c4-41ab-944f-9e9507afe1a0","Type":"ContainerStarted","Data":"62e0d64b23bf6963dc5d2698298fab9a9aca96a45bb5b55b6508cfcc3d4ecf22"} Dec 08 15:39:50 crc kubenswrapper[4894]: I1208 15:39:50.634712 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4z2f" event={"ID":"e5806d93-34c4-41ab-944f-9e9507afe1a0","Type":"ContainerStarted","Data":"3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e"} Dec 08 15:39:51 crc kubenswrapper[4894]: I1208 15:39:51.197052 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:39:51 crc kubenswrapper[4894]: E1208 15:39:51.197412 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:39:51 crc kubenswrapper[4894]: I1208 15:39:51.644936 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerID="3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e" exitCode=0 Dec 08 15:39:51 crc kubenswrapper[4894]: I1208 15:39:51.644980 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4z2f" event={"ID":"e5806d93-34c4-41ab-944f-9e9507afe1a0","Type":"ContainerDied","Data":"3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e"} Dec 08 15:39:52 crc kubenswrapper[4894]: I1208 15:39:52.655406 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4z2f" event={"ID":"e5806d93-34c4-41ab-944f-9e9507afe1a0","Type":"ContainerStarted","Data":"40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f"} Dec 08 15:39:52 crc kubenswrapper[4894]: I1208 15:39:52.684157 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c4z2f" podStartSLOduration=3.158371329 podStartE2EDuration="5.684127025s" podCreationTimestamp="2025-12-08 15:39:47 +0000 UTC" firstStartedPulling="2025-12-08 15:39:49.626772317 +0000 UTC m=+3210.726778432" lastFinishedPulling="2025-12-08 15:39:52.152528013 +0000 UTC m=+3213.252534128" observedRunningTime="2025-12-08 15:39:52.674189014 +0000 UTC m=+3213.774195129" watchObservedRunningTime="2025-12-08 15:39:52.684127025 +0000 UTC m=+3213.784133140" Dec 08 15:39:58 crc kubenswrapper[4894]: I1208 15:39:58.249715 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:58 crc kubenswrapper[4894]: I1208 15:39:58.250274 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:58 crc kubenswrapper[4894]: I1208 15:39:58.337501 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:58 crc kubenswrapper[4894]: I1208 15:39:58.778065 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:39:58 crc kubenswrapper[4894]: I1208 15:39:58.836332 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4z2f"] Dec 08 15:40:00 crc kubenswrapper[4894]: I1208 15:40:00.721102 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c4z2f" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="registry-server" containerID="cri-o://40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f" gracePeriod=2 Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.212676 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.317528 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szrxr\" (UniqueName: \"kubernetes.io/projected/e5806d93-34c4-41ab-944f-9e9507afe1a0-kube-api-access-szrxr\") pod \"e5806d93-34c4-41ab-944f-9e9507afe1a0\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.317604 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-utilities\") pod \"e5806d93-34c4-41ab-944f-9e9507afe1a0\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.317664 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-catalog-content\") pod \"e5806d93-34c4-41ab-944f-9e9507afe1a0\" (UID: \"e5806d93-34c4-41ab-944f-9e9507afe1a0\") " Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.318504 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-utilities" (OuterVolumeSpecName: "utilities") pod "e5806d93-34c4-41ab-944f-9e9507afe1a0" (UID: "e5806d93-34c4-41ab-944f-9e9507afe1a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.324888 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5806d93-34c4-41ab-944f-9e9507afe1a0-kube-api-access-szrxr" (OuterVolumeSpecName: "kube-api-access-szrxr") pod "e5806d93-34c4-41ab-944f-9e9507afe1a0" (UID: "e5806d93-34c4-41ab-944f-9e9507afe1a0"). InnerVolumeSpecName "kube-api-access-szrxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.376520 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5806d93-34c4-41ab-944f-9e9507afe1a0" (UID: "e5806d93-34c4-41ab-944f-9e9507afe1a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.420523 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szrxr\" (UniqueName: \"kubernetes.io/projected/e5806d93-34c4-41ab-944f-9e9507afe1a0-kube-api-access-szrxr\") on node \"crc\" DevicePath \"\"" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.420554 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.420564 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5806d93-34c4-41ab-944f-9e9507afe1a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.730085 4894 generic.go:334] "Generic (PLEG): container finished" podID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerID="40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f" exitCode=0 Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.730130 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4z2f" event={"ID":"e5806d93-34c4-41ab-944f-9e9507afe1a0","Type":"ContainerDied","Data":"40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f"} Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.730158 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4z2f" event={"ID":"e5806d93-34c4-41ab-944f-9e9507afe1a0","Type":"ContainerDied","Data":"62e0d64b23bf6963dc5d2698298fab9a9aca96a45bb5b55b6508cfcc3d4ecf22"} Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.730175 4894 scope.go:117] "RemoveContainer" containerID="40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.730307 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4z2f" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.755237 4894 scope.go:117] "RemoveContainer" containerID="3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.774189 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4z2f"] Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.779240 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c4z2f"] Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.787134 4894 scope.go:117] "RemoveContainer" containerID="105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.846722 4894 scope.go:117] "RemoveContainer" containerID="40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f" Dec 08 15:40:01 crc kubenswrapper[4894]: E1208 15:40:01.847182 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f\": container with ID starting with 40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f not found: ID does not exist" containerID="40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.847224 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f"} err="failed to get container status \"40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f\": rpc error: code = NotFound desc = could not find container \"40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f\": container with ID starting with 40db8ba00eee1a12b3a391bc93f1ac158b2ed928a8a18f784dd5f61286a46f1f not found: ID does not exist" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.847250 4894 scope.go:117] "RemoveContainer" containerID="3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e" Dec 08 15:40:01 crc kubenswrapper[4894]: E1208 15:40:01.847732 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e\": container with ID starting with 3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e not found: ID does not exist" containerID="3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.847758 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e"} err="failed to get container status \"3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e\": rpc error: code = NotFound desc = could not find container \"3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e\": container with ID starting with 3748d92223c9a78b2729b2918b522f55118d24003b1c95cc89d91cc72f1dc55e not found: ID does not exist" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.847776 4894 scope.go:117] "RemoveContainer" containerID="105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d" Dec 08 15:40:01 crc kubenswrapper[4894]: E1208 15:40:01.848147 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d\": container with ID starting with 105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d not found: ID does not exist" containerID="105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d" Dec 08 15:40:01 crc kubenswrapper[4894]: I1208 15:40:01.848181 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d"} err="failed to get container status \"105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d\": rpc error: code = NotFound desc = could not find container \"105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d\": container with ID starting with 105e98f8afaa1f10eb38b8e66515b33e6ce656986decd10bdd15c8201530f48d not found: ID does not exist" Dec 08 15:40:03 crc kubenswrapper[4894]: I1208 15:40:03.197573 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:40:03 crc kubenswrapper[4894]: E1208 15:40:03.197864 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:40:03 crc kubenswrapper[4894]: I1208 15:40:03.215641 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" path="/var/lib/kubelet/pods/e5806d93-34c4-41ab-944f-9e9507afe1a0/volumes" Dec 08 15:40:15 crc kubenswrapper[4894]: I1208 15:40:15.197589 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:40:15 crc kubenswrapper[4894]: I1208 15:40:15.914411 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"65921f34e15335690428cdddc29012738b164fb0aea2b952bd02813b0a2dad6e"} Dec 08 15:42:22 crc kubenswrapper[4894]: I1208 15:42:22.124023 4894 generic.go:334] "Generic (PLEG): container finished" podID="39de6187-9721-4687-8001-ef13c4abafb3" containerID="71f05a5c05e9306bfbe7afecb831d54408777c5adddd4df2b12f38584c049d2e" exitCode=0 Dec 08 15:42:22 crc kubenswrapper[4894]: I1208 15:42:22.124105 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"39de6187-9721-4687-8001-ef13c4abafb3","Type":"ContainerDied","Data":"71f05a5c05e9306bfbe7afecb831d54408777c5adddd4df2b12f38584c049d2e"} Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.624308 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797155 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-workdir\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797245 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797267 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ca-certs\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797290 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-temporary\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797326 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl6bp\" (UniqueName: \"kubernetes.io/projected/39de6187-9721-4687-8001-ef13c4abafb3-kube-api-access-kl6bp\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797429 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config-secret\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797504 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-config-data\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797575 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ssh-key\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.797608 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"39de6187-9721-4687-8001-ef13c4abafb3\" (UID: \"39de6187-9721-4687-8001-ef13c4abafb3\") " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.798627 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-config-data" (OuterVolumeSpecName: "config-data") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.799209 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.803070 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.803555 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39de6187-9721-4687-8001-ef13c4abafb3-kube-api-access-kl6bp" (OuterVolumeSpecName: "kube-api-access-kl6bp") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "kube-api-access-kl6bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.809696 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.826691 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.827531 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.832684 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.850145 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "39de6187-9721-4687-8001-ef13c4abafb3" (UID: "39de6187-9721-4687-8001-ef13c4abafb3"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899455 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899494 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899508 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899520 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/39de6187-9721-4687-8001-ef13c4abafb3-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899532 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl6bp\" (UniqueName: \"kubernetes.io/projected/39de6187-9721-4687-8001-ef13c4abafb3-kube-api-access-kl6bp\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899543 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899552 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39de6187-9721-4687-8001-ef13c4abafb3-config-data\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899561 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39de6187-9721-4687-8001-ef13c4abafb3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.899604 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 08 15:42:23 crc kubenswrapper[4894]: I1208 15:42:23.919461 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 08 15:42:24 crc kubenswrapper[4894]: I1208 15:42:24.001904 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:24 crc kubenswrapper[4894]: I1208 15:42:24.141401 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"39de6187-9721-4687-8001-ef13c4abafb3","Type":"ContainerDied","Data":"974b3d602f8a19bc8dd6f4c96523d145d71f64fbab0b98a313a891796443b5a4"} Dec 08 15:42:24 crc kubenswrapper[4894]: I1208 15:42:24.141442 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974b3d602f8a19bc8dd6f4c96523d145d71f64fbab0b98a313a891796443b5a4" Dec 08 15:42:24 crc kubenswrapper[4894]: I1208 15:42:24.141679 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.162122 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mph96"] Dec 08 15:42:27 crc kubenswrapper[4894]: E1208 15:42:27.162921 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39de6187-9721-4687-8001-ef13c4abafb3" containerName="tempest-tests-tempest-tests-runner" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.162934 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="39de6187-9721-4687-8001-ef13c4abafb3" containerName="tempest-tests-tempest-tests-runner" Dec 08 15:42:27 crc kubenswrapper[4894]: E1208 15:42:27.162949 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="extract-utilities" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.162956 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="extract-utilities" Dec 08 15:42:27 crc kubenswrapper[4894]: E1208 15:42:27.162969 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="registry-server" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.162975 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="registry-server" Dec 08 15:42:27 crc kubenswrapper[4894]: E1208 15:42:27.162998 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="extract-content" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.163004 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="extract-content" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.163175 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5806d93-34c4-41ab-944f-9e9507afe1a0" containerName="registry-server" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.163198 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="39de6187-9721-4687-8001-ef13c4abafb3" containerName="tempest-tests-tempest-tests-runner" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.164449 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.191421 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mph96"] Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.266022 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-catalog-content\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.267286 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pqng\" (UniqueName: \"kubernetes.io/projected/bd527363-d42e-4f9f-9132-b8f5babea1a6-kube-api-access-2pqng\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.267405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-utilities\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.369922 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pqng\" (UniqueName: \"kubernetes.io/projected/bd527363-d42e-4f9f-9132-b8f5babea1a6-kube-api-access-2pqng\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.370042 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-utilities\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.370131 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-catalog-content\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.370834 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-utilities\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.370909 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-catalog-content\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.388600 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pqng\" (UniqueName: \"kubernetes.io/projected/bd527363-d42e-4f9f-9132-b8f5babea1a6-kube-api-access-2pqng\") pod \"community-operators-mph96\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.496313 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:27 crc kubenswrapper[4894]: I1208 15:42:27.984159 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mph96"] Dec 08 15:42:28 crc kubenswrapper[4894]: I1208 15:42:28.182000 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerStarted","Data":"32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a"} Dec 08 15:42:28 crc kubenswrapper[4894]: I1208 15:42:28.182036 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerStarted","Data":"5a0b4e818e4145347ce89b2287484154cb8f05534a3e61696c63c9054baeed81"} Dec 08 15:42:28 crc kubenswrapper[4894]: I1208 15:42:28.185168 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:42:29 crc kubenswrapper[4894]: I1208 15:42:29.195320 4894 generic.go:334] "Generic (PLEG): container finished" podID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerID="32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a" exitCode=0 Dec 08 15:42:29 crc kubenswrapper[4894]: I1208 15:42:29.195489 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerDied","Data":"32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a"} Dec 08 15:42:29 crc kubenswrapper[4894]: I1208 15:42:29.195770 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerStarted","Data":"1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea"} Dec 08 15:42:30 crc kubenswrapper[4894]: I1208 15:42:30.210075 4894 generic.go:334] "Generic (PLEG): container finished" podID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerID="1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea" exitCode=0 Dec 08 15:42:30 crc kubenswrapper[4894]: I1208 15:42:30.210185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerDied","Data":"1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea"} Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.222369 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerStarted","Data":"5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7"} Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.764104 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mph96" podStartSLOduration=2.355619908 podStartE2EDuration="4.764078567s" podCreationTimestamp="2025-12-08 15:42:27 +0000 UTC" firstStartedPulling="2025-12-08 15:42:28.184718837 +0000 UTC m=+3369.284724952" lastFinishedPulling="2025-12-08 15:42:30.593177456 +0000 UTC m=+3371.693183611" observedRunningTime="2025-12-08 15:42:31.253858104 +0000 UTC m=+3372.353864219" watchObservedRunningTime="2025-12-08 15:42:31.764078567 +0000 UTC m=+3372.864084682" Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.766721 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.768073 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.770024 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-dq2lk" Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.777649 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.946518 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:31 crc kubenswrapper[4894]: I1208 15:42:31.946665 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q465t\" (UniqueName: \"kubernetes.io/projected/1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb-kube-api-access-q465t\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:32 crc kubenswrapper[4894]: I1208 15:42:32.048731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q465t\" (UniqueName: \"kubernetes.io/projected/1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb-kube-api-access-q465t\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:32 crc kubenswrapper[4894]: I1208 15:42:32.049119 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:32 crc kubenswrapper[4894]: I1208 15:42:32.049578 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:32 crc kubenswrapper[4894]: I1208 15:42:32.069400 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q465t\" (UniqueName: \"kubernetes.io/projected/1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb-kube-api-access-q465t\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:32 crc kubenswrapper[4894]: I1208 15:42:32.084302 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:32 crc kubenswrapper[4894]: I1208 15:42:32.089779 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 08 15:42:32 crc kubenswrapper[4894]: I1208 15:42:32.546710 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 08 15:42:32 crc kubenswrapper[4894]: W1208 15:42:32.548144 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b11a307_3d6a_4f5e_82eb_b864ef5ed5cb.slice/crio-850103b563adf54a0e84a302c5e5df9d2b9e7b8079410e1a85f02330aa29ec31 WatchSource:0}: Error finding container 850103b563adf54a0e84a302c5e5df9d2b9e7b8079410e1a85f02330aa29ec31: Status 404 returned error can't find the container with id 850103b563adf54a0e84a302c5e5df9d2b9e7b8079410e1a85f02330aa29ec31 Dec 08 15:42:33 crc kubenswrapper[4894]: I1208 15:42:33.242138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb","Type":"ContainerStarted","Data":"850103b563adf54a0e84a302c5e5df9d2b9e7b8079410e1a85f02330aa29ec31"} Dec 08 15:42:36 crc kubenswrapper[4894]: I1208 15:42:36.265534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb","Type":"ContainerStarted","Data":"7e38909027d81ab0fc42d6dcd7fbbb2f6252083bd53697e3be7de0018760711c"} Dec 08 15:42:36 crc kubenswrapper[4894]: I1208 15:42:36.280163 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.4045540819999998 podStartE2EDuration="5.280146136s" podCreationTimestamp="2025-12-08 15:42:31 +0000 UTC" firstStartedPulling="2025-12-08 15:42:32.550160637 +0000 UTC m=+3373.650166752" lastFinishedPulling="2025-12-08 15:42:35.425752691 +0000 UTC m=+3376.525758806" observedRunningTime="2025-12-08 15:42:36.277030339 +0000 UTC m=+3377.377036454" watchObservedRunningTime="2025-12-08 15:42:36.280146136 +0000 UTC m=+3377.380152251" Dec 08 15:42:37 crc kubenswrapper[4894]: I1208 15:42:37.296415 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:42:37 crc kubenswrapper[4894]: I1208 15:42:37.296474 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:42:37 crc kubenswrapper[4894]: I1208 15:42:37.496545 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:37 crc kubenswrapper[4894]: I1208 15:42:37.496594 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:37 crc kubenswrapper[4894]: I1208 15:42:37.561430 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:38 crc kubenswrapper[4894]: I1208 15:42:38.337754 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:38 crc kubenswrapper[4894]: I1208 15:42:38.384688 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mph96"] Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.296844 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mph96" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="registry-server" containerID="cri-o://5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7" gracePeriod=2 Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.779600 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.914724 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-catalog-content\") pod \"bd527363-d42e-4f9f-9132-b8f5babea1a6\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.914770 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-utilities\") pod \"bd527363-d42e-4f9f-9132-b8f5babea1a6\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.914895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pqng\" (UniqueName: \"kubernetes.io/projected/bd527363-d42e-4f9f-9132-b8f5babea1a6-kube-api-access-2pqng\") pod \"bd527363-d42e-4f9f-9132-b8f5babea1a6\" (UID: \"bd527363-d42e-4f9f-9132-b8f5babea1a6\") " Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.916461 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-utilities" (OuterVolumeSpecName: "utilities") pod "bd527363-d42e-4f9f-9132-b8f5babea1a6" (UID: "bd527363-d42e-4f9f-9132-b8f5babea1a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.921503 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd527363-d42e-4f9f-9132-b8f5babea1a6-kube-api-access-2pqng" (OuterVolumeSpecName: "kube-api-access-2pqng") pod "bd527363-d42e-4f9f-9132-b8f5babea1a6" (UID: "bd527363-d42e-4f9f-9132-b8f5babea1a6"). InnerVolumeSpecName "kube-api-access-2pqng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:42:40 crc kubenswrapper[4894]: I1208 15:42:40.970764 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd527363-d42e-4f9f-9132-b8f5babea1a6" (UID: "bd527363-d42e-4f9f-9132-b8f5babea1a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.016574 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pqng\" (UniqueName: \"kubernetes.io/projected/bd527363-d42e-4f9f-9132-b8f5babea1a6-kube-api-access-2pqng\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.016600 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.016609 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd527363-d42e-4f9f-9132-b8f5babea1a6-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.308195 4894 generic.go:334] "Generic (PLEG): container finished" podID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerID="5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7" exitCode=0 Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.309095 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerDied","Data":"5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7"} Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.309186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mph96" event={"ID":"bd527363-d42e-4f9f-9132-b8f5babea1a6","Type":"ContainerDied","Data":"5a0b4e818e4145347ce89b2287484154cb8f05534a3e61696c63c9054baeed81"} Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.309267 4894 scope.go:117] "RemoveContainer" containerID="5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.309462 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mph96" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.335068 4894 scope.go:117] "RemoveContainer" containerID="1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.341342 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mph96"] Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.349956 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mph96"] Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.356000 4894 scope.go:117] "RemoveContainer" containerID="32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.398331 4894 scope.go:117] "RemoveContainer" containerID="5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7" Dec 08 15:42:41 crc kubenswrapper[4894]: E1208 15:42:41.398757 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7\": container with ID starting with 5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7 not found: ID does not exist" containerID="5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.398789 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7"} err="failed to get container status \"5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7\": rpc error: code = NotFound desc = could not find container \"5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7\": container with ID starting with 5abdbf2a6858c77a9c3cd81b35a1c5d1c84835c2a34ae1414ba6705dab209ef7 not found: ID does not exist" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.398810 4894 scope.go:117] "RemoveContainer" containerID="1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea" Dec 08 15:42:41 crc kubenswrapper[4894]: E1208 15:42:41.399190 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea\": container with ID starting with 1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea not found: ID does not exist" containerID="1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.399215 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea"} err="failed to get container status \"1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea\": rpc error: code = NotFound desc = could not find container \"1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea\": container with ID starting with 1edd2b680840506ab668b92d76561daef0bd9894427cba25a068d8624fe1dfea not found: ID does not exist" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.399229 4894 scope.go:117] "RemoveContainer" containerID="32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a" Dec 08 15:42:41 crc kubenswrapper[4894]: E1208 15:42:41.399500 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a\": container with ID starting with 32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a not found: ID does not exist" containerID="32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a" Dec 08 15:42:41 crc kubenswrapper[4894]: I1208 15:42:41.399522 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a"} err="failed to get container status \"32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a\": rpc error: code = NotFound desc = could not find container \"32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a\": container with ID starting with 32ebe4b532888c078ac23654f104b9bd3108bd75ca2818fe7cd6fd605877ce6a not found: ID does not exist" Dec 08 15:42:43 crc kubenswrapper[4894]: I1208 15:42:43.207101 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" path="/var/lib/kubelet/pods/bd527363-d42e-4f9f-9132-b8f5babea1a6/volumes" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.425220 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c4b76/must-gather-fjsbr"] Dec 08 15:42:57 crc kubenswrapper[4894]: E1208 15:42:57.426145 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="extract-utilities" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.426158 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="extract-utilities" Dec 08 15:42:57 crc kubenswrapper[4894]: E1208 15:42:57.426176 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="extract-content" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.426183 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="extract-content" Dec 08 15:42:57 crc kubenswrapper[4894]: E1208 15:42:57.426193 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="registry-server" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.426199 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="registry-server" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.426366 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd527363-d42e-4f9f-9132-b8f5babea1a6" containerName="registry-server" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.427406 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.430157 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-c4b76"/"default-dockercfg-jgz74" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.430245 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c4b76"/"kube-root-ca.crt" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.430898 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c4b76"/"openshift-service-ca.crt" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.435489 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c4b76/must-gather-fjsbr"] Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.530528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfr5w\" (UniqueName: \"kubernetes.io/projected/23675181-ae3e-4a38-9f92-06ed475dfa33-kube-api-access-vfr5w\") pod \"must-gather-fjsbr\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.530699 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23675181-ae3e-4a38-9f92-06ed475dfa33-must-gather-output\") pod \"must-gather-fjsbr\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.632056 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfr5w\" (UniqueName: \"kubernetes.io/projected/23675181-ae3e-4a38-9f92-06ed475dfa33-kube-api-access-vfr5w\") pod \"must-gather-fjsbr\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.632224 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23675181-ae3e-4a38-9f92-06ed475dfa33-must-gather-output\") pod \"must-gather-fjsbr\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.632696 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23675181-ae3e-4a38-9f92-06ed475dfa33-must-gather-output\") pod \"must-gather-fjsbr\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.650669 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfr5w\" (UniqueName: \"kubernetes.io/projected/23675181-ae3e-4a38-9f92-06ed475dfa33-kube-api-access-vfr5w\") pod \"must-gather-fjsbr\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:57 crc kubenswrapper[4894]: I1208 15:42:57.751446 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:42:58 crc kubenswrapper[4894]: I1208 15:42:58.197436 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c4b76/must-gather-fjsbr"] Dec 08 15:42:58 crc kubenswrapper[4894]: W1208 15:42:58.200992 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23675181_ae3e_4a38_9f92_06ed475dfa33.slice/crio-c294549b4ac9d13d28ed3d766d87d9700350baeb6dcbcac5c41c4e519765c41a WatchSource:0}: Error finding container c294549b4ac9d13d28ed3d766d87d9700350baeb6dcbcac5c41c4e519765c41a: Status 404 returned error can't find the container with id c294549b4ac9d13d28ed3d766d87d9700350baeb6dcbcac5c41c4e519765c41a Dec 08 15:42:58 crc kubenswrapper[4894]: I1208 15:42:58.508957 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/must-gather-fjsbr" event={"ID":"23675181-ae3e-4a38-9f92-06ed475dfa33","Type":"ContainerStarted","Data":"c294549b4ac9d13d28ed3d766d87d9700350baeb6dcbcac5c41c4e519765c41a"} Dec 08 15:43:05 crc kubenswrapper[4894]: I1208 15:43:05.581435 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/must-gather-fjsbr" event={"ID":"23675181-ae3e-4a38-9f92-06ed475dfa33","Type":"ContainerStarted","Data":"575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662"} Dec 08 15:43:05 crc kubenswrapper[4894]: I1208 15:43:05.581928 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/must-gather-fjsbr" event={"ID":"23675181-ae3e-4a38-9f92-06ed475dfa33","Type":"ContainerStarted","Data":"ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa"} Dec 08 15:43:05 crc kubenswrapper[4894]: I1208 15:43:05.606463 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c4b76/must-gather-fjsbr" podStartSLOduration=1.749988836 podStartE2EDuration="8.606445214s" podCreationTimestamp="2025-12-08 15:42:57 +0000 UTC" firstStartedPulling="2025-12-08 15:42:58.203004733 +0000 UTC m=+3399.303010848" lastFinishedPulling="2025-12-08 15:43:05.059461111 +0000 UTC m=+3406.159467226" observedRunningTime="2025-12-08 15:43:05.596655238 +0000 UTC m=+3406.696661353" watchObservedRunningTime="2025-12-08 15:43:05.606445214 +0000 UTC m=+3406.706451329" Dec 08 15:43:07 crc kubenswrapper[4894]: I1208 15:43:07.297212 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:43:07 crc kubenswrapper[4894]: I1208 15:43:07.297538 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:43:08 crc kubenswrapper[4894]: I1208 15:43:08.939585 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c4b76/crc-debug-kwrvv"] Dec 08 15:43:08 crc kubenswrapper[4894]: I1208 15:43:08.941320 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.069304 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvm5q\" (UniqueName: \"kubernetes.io/projected/3793e2a4-6435-4b44-a610-9ec65c5967a5-kube-api-access-dvm5q\") pod \"crc-debug-kwrvv\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.069495 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3793e2a4-6435-4b44-a610-9ec65c5967a5-host\") pod \"crc-debug-kwrvv\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.170834 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvm5q\" (UniqueName: \"kubernetes.io/projected/3793e2a4-6435-4b44-a610-9ec65c5967a5-kube-api-access-dvm5q\") pod \"crc-debug-kwrvv\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.171030 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3793e2a4-6435-4b44-a610-9ec65c5967a5-host\") pod \"crc-debug-kwrvv\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.171149 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3793e2a4-6435-4b44-a610-9ec65c5967a5-host\") pod \"crc-debug-kwrvv\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.191473 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvm5q\" (UniqueName: \"kubernetes.io/projected/3793e2a4-6435-4b44-a610-9ec65c5967a5-kube-api-access-dvm5q\") pod \"crc-debug-kwrvv\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.256185 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:43:09 crc kubenswrapper[4894]: W1208 15:43:09.488015 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3793e2a4_6435_4b44_a610_9ec65c5967a5.slice/crio-20ece7b1f6f631931f74aaba6abc49a84102e55e74370c305fd16666e462eae7 WatchSource:0}: Error finding container 20ece7b1f6f631931f74aaba6abc49a84102e55e74370c305fd16666e462eae7: Status 404 returned error can't find the container with id 20ece7b1f6f631931f74aaba6abc49a84102e55e74370c305fd16666e462eae7 Dec 08 15:43:09 crc kubenswrapper[4894]: I1208 15:43:09.617512 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" event={"ID":"3793e2a4-6435-4b44-a610-9ec65c5967a5","Type":"ContainerStarted","Data":"20ece7b1f6f631931f74aaba6abc49a84102e55e74370c305fd16666e462eae7"} Dec 08 15:43:20 crc kubenswrapper[4894]: I1208 15:43:20.729063 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" event={"ID":"3793e2a4-6435-4b44-a610-9ec65c5967a5","Type":"ContainerStarted","Data":"7ff7270ae40210262218f106b0abc030182b20dc5e46072ecf0607300cd59c48"} Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.297027 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.297612 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.297662 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.298448 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65921f34e15335690428cdddc29012738b164fb0aea2b952bd02813b0a2dad6e"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.298514 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://65921f34e15335690428cdddc29012738b164fb0aea2b952bd02813b0a2dad6e" gracePeriod=600 Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.873139 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="65921f34e15335690428cdddc29012738b164fb0aea2b952bd02813b0a2dad6e" exitCode=0 Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.873289 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"65921f34e15335690428cdddc29012738b164fb0aea2b952bd02813b0a2dad6e"} Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.873742 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295"} Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.873761 4894 scope.go:117] "RemoveContainer" containerID="535cc58686052fec9fdbd6a4c7192266f3692f4b38d3c8d2946e9516fffa54a8" Dec 08 15:43:37 crc kubenswrapper[4894]: I1208 15:43:37.899208 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" podStartSLOduration=19.033386334 podStartE2EDuration="29.899189087s" podCreationTimestamp="2025-12-08 15:43:08 +0000 UTC" firstStartedPulling="2025-12-08 15:43:09.498047697 +0000 UTC m=+3410.598053812" lastFinishedPulling="2025-12-08 15:43:20.36385045 +0000 UTC m=+3421.463856565" observedRunningTime="2025-12-08 15:43:20.748305681 +0000 UTC m=+3421.848311796" watchObservedRunningTime="2025-12-08 15:43:37.899189087 +0000 UTC m=+3438.999195202" Dec 08 15:44:00 crc kubenswrapper[4894]: I1208 15:44:00.083561 4894 generic.go:334] "Generic (PLEG): container finished" podID="3793e2a4-6435-4b44-a610-9ec65c5967a5" containerID="7ff7270ae40210262218f106b0abc030182b20dc5e46072ecf0607300cd59c48" exitCode=0 Dec 08 15:44:00 crc kubenswrapper[4894]: I1208 15:44:00.083665 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" event={"ID":"3793e2a4-6435-4b44-a610-9ec65c5967a5","Type":"ContainerDied","Data":"7ff7270ae40210262218f106b0abc030182b20dc5e46072ecf0607300cd59c48"} Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.196197 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.251374 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c4b76/crc-debug-kwrvv"] Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.261087 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c4b76/crc-debug-kwrvv"] Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.314948 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3793e2a4-6435-4b44-a610-9ec65c5967a5-host\") pod \"3793e2a4-6435-4b44-a610-9ec65c5967a5\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.315088 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3793e2a4-6435-4b44-a610-9ec65c5967a5-host" (OuterVolumeSpecName: "host") pod "3793e2a4-6435-4b44-a610-9ec65c5967a5" (UID: "3793e2a4-6435-4b44-a610-9ec65c5967a5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.315195 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvm5q\" (UniqueName: \"kubernetes.io/projected/3793e2a4-6435-4b44-a610-9ec65c5967a5-kube-api-access-dvm5q\") pod \"3793e2a4-6435-4b44-a610-9ec65c5967a5\" (UID: \"3793e2a4-6435-4b44-a610-9ec65c5967a5\") " Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.315713 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3793e2a4-6435-4b44-a610-9ec65c5967a5-host\") on node \"crc\" DevicePath \"\"" Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.325082 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3793e2a4-6435-4b44-a610-9ec65c5967a5-kube-api-access-dvm5q" (OuterVolumeSpecName: "kube-api-access-dvm5q") pod "3793e2a4-6435-4b44-a610-9ec65c5967a5" (UID: "3793e2a4-6435-4b44-a610-9ec65c5967a5"). InnerVolumeSpecName "kube-api-access-dvm5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:44:01 crc kubenswrapper[4894]: I1208 15:44:01.417876 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvm5q\" (UniqueName: \"kubernetes.io/projected/3793e2a4-6435-4b44-a610-9ec65c5967a5-kube-api-access-dvm5q\") on node \"crc\" DevicePath \"\"" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.105144 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20ece7b1f6f631931f74aaba6abc49a84102e55e74370c305fd16666e462eae7" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.105228 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-kwrvv" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.467908 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c4b76/crc-debug-pj86q"] Dec 08 15:44:02 crc kubenswrapper[4894]: E1208 15:44:02.468303 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3793e2a4-6435-4b44-a610-9ec65c5967a5" containerName="container-00" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.468316 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3793e2a4-6435-4b44-a610-9ec65c5967a5" containerName="container-00" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.468505 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3793e2a4-6435-4b44-a610-9ec65c5967a5" containerName="container-00" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.469205 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.543801 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58adb9cc-ed4d-4d75-863e-e09680757175-host\") pod \"crc-debug-pj86q\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.544180 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdhcf\" (UniqueName: \"kubernetes.io/projected/58adb9cc-ed4d-4d75-863e-e09680757175-kube-api-access-wdhcf\") pod \"crc-debug-pj86q\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.646986 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdhcf\" (UniqueName: \"kubernetes.io/projected/58adb9cc-ed4d-4d75-863e-e09680757175-kube-api-access-wdhcf\") pod \"crc-debug-pj86q\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.647207 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58adb9cc-ed4d-4d75-863e-e09680757175-host\") pod \"crc-debug-pj86q\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.647436 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58adb9cc-ed4d-4d75-863e-e09680757175-host\") pod \"crc-debug-pj86q\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.672277 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdhcf\" (UniqueName: \"kubernetes.io/projected/58adb9cc-ed4d-4d75-863e-e09680757175-kube-api-access-wdhcf\") pod \"crc-debug-pj86q\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:02 crc kubenswrapper[4894]: I1208 15:44:02.795881 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:03 crc kubenswrapper[4894]: I1208 15:44:03.117006 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/crc-debug-pj86q" event={"ID":"58adb9cc-ed4d-4d75-863e-e09680757175","Type":"ContainerStarted","Data":"8e495136b16762fc0fa2111e9f96b6e849309dd7924c67b37f73289ee638b02a"} Dec 08 15:44:03 crc kubenswrapper[4894]: I1208 15:44:03.208186 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3793e2a4-6435-4b44-a610-9ec65c5967a5" path="/var/lib/kubelet/pods/3793e2a4-6435-4b44-a610-9ec65c5967a5/volumes" Dec 08 15:44:04 crc kubenswrapper[4894]: I1208 15:44:04.127787 4894 generic.go:334] "Generic (PLEG): container finished" podID="58adb9cc-ed4d-4d75-863e-e09680757175" containerID="c8a0f6f74edecfe1d54fe5d6e89ea33118239abe3498e1799800c8b8ea4c2d7c" exitCode=0 Dec 08 15:44:04 crc kubenswrapper[4894]: I1208 15:44:04.127874 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/crc-debug-pj86q" event={"ID":"58adb9cc-ed4d-4d75-863e-e09680757175","Type":"ContainerDied","Data":"c8a0f6f74edecfe1d54fe5d6e89ea33118239abe3498e1799800c8b8ea4c2d7c"} Dec 08 15:44:04 crc kubenswrapper[4894]: I1208 15:44:04.664447 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c4b76/crc-debug-pj86q"] Dec 08 15:44:04 crc kubenswrapper[4894]: I1208 15:44:04.672433 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c4b76/crc-debug-pj86q"] Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.253091 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.400482 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58adb9cc-ed4d-4d75-863e-e09680757175-host\") pod \"58adb9cc-ed4d-4d75-863e-e09680757175\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.400608 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58adb9cc-ed4d-4d75-863e-e09680757175-host" (OuterVolumeSpecName: "host") pod "58adb9cc-ed4d-4d75-863e-e09680757175" (UID: "58adb9cc-ed4d-4d75-863e-e09680757175"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.400900 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdhcf\" (UniqueName: \"kubernetes.io/projected/58adb9cc-ed4d-4d75-863e-e09680757175-kube-api-access-wdhcf\") pod \"58adb9cc-ed4d-4d75-863e-e09680757175\" (UID: \"58adb9cc-ed4d-4d75-863e-e09680757175\") " Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.401310 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58adb9cc-ed4d-4d75-863e-e09680757175-host\") on node \"crc\" DevicePath \"\"" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.405393 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58adb9cc-ed4d-4d75-863e-e09680757175-kube-api-access-wdhcf" (OuterVolumeSpecName: "kube-api-access-wdhcf") pod "58adb9cc-ed4d-4d75-863e-e09680757175" (UID: "58adb9cc-ed4d-4d75-863e-e09680757175"). InnerVolumeSpecName "kube-api-access-wdhcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.502839 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdhcf\" (UniqueName: \"kubernetes.io/projected/58adb9cc-ed4d-4d75-863e-e09680757175-kube-api-access-wdhcf\") on node \"crc\" DevicePath \"\"" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.837641 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c4b76/crc-debug-5lxbm"] Dec 08 15:44:05 crc kubenswrapper[4894]: E1208 15:44:05.838108 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58adb9cc-ed4d-4d75-863e-e09680757175" containerName="container-00" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.838133 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="58adb9cc-ed4d-4d75-863e-e09680757175" containerName="container-00" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.838322 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="58adb9cc-ed4d-4d75-863e-e09680757175" containerName="container-00" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.838968 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.910060 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29zps\" (UniqueName: \"kubernetes.io/projected/7530a157-fc6c-4207-9725-89f3b2d2806b-kube-api-access-29zps\") pod \"crc-debug-5lxbm\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:05 crc kubenswrapper[4894]: I1208 15:44:05.910112 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530a157-fc6c-4207-9725-89f3b2d2806b-host\") pod \"crc-debug-5lxbm\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:06 crc kubenswrapper[4894]: I1208 15:44:06.011539 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29zps\" (UniqueName: \"kubernetes.io/projected/7530a157-fc6c-4207-9725-89f3b2d2806b-kube-api-access-29zps\") pod \"crc-debug-5lxbm\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:06 crc kubenswrapper[4894]: I1208 15:44:06.011896 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530a157-fc6c-4207-9725-89f3b2d2806b-host\") pod \"crc-debug-5lxbm\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:06 crc kubenswrapper[4894]: I1208 15:44:06.011976 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530a157-fc6c-4207-9725-89f3b2d2806b-host\") pod \"crc-debug-5lxbm\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:06 crc kubenswrapper[4894]: I1208 15:44:06.027524 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29zps\" (UniqueName: \"kubernetes.io/projected/7530a157-fc6c-4207-9725-89f3b2d2806b-kube-api-access-29zps\") pod \"crc-debug-5lxbm\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:06 crc kubenswrapper[4894]: I1208 15:44:06.147017 4894 scope.go:117] "RemoveContainer" containerID="c8a0f6f74edecfe1d54fe5d6e89ea33118239abe3498e1799800c8b8ea4c2d7c" Dec 08 15:44:06 crc kubenswrapper[4894]: I1208 15:44:06.147049 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-pj86q" Dec 08 15:44:06 crc kubenswrapper[4894]: I1208 15:44:06.155142 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:06 crc kubenswrapper[4894]: W1208 15:44:06.204454 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7530a157_fc6c_4207_9725_89f3b2d2806b.slice/crio-e2ac402ea256b79cce48c399c07730f98fe1b4de89f6828860449d6f92fb896e WatchSource:0}: Error finding container e2ac402ea256b79cce48c399c07730f98fe1b4de89f6828860449d6f92fb896e: Status 404 returned error can't find the container with id e2ac402ea256b79cce48c399c07730f98fe1b4de89f6828860449d6f92fb896e Dec 08 15:44:07 crc kubenswrapper[4894]: I1208 15:44:07.158034 4894 generic.go:334] "Generic (PLEG): container finished" podID="7530a157-fc6c-4207-9725-89f3b2d2806b" containerID="a3c64c0e829a54a4a585eeb857858b820ededc4925c8d04836d8f1c798796521" exitCode=0 Dec 08 15:44:07 crc kubenswrapper[4894]: I1208 15:44:07.158101 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/crc-debug-5lxbm" event={"ID":"7530a157-fc6c-4207-9725-89f3b2d2806b","Type":"ContainerDied","Data":"a3c64c0e829a54a4a585eeb857858b820ededc4925c8d04836d8f1c798796521"} Dec 08 15:44:07 crc kubenswrapper[4894]: I1208 15:44:07.158493 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/crc-debug-5lxbm" event={"ID":"7530a157-fc6c-4207-9725-89f3b2d2806b","Type":"ContainerStarted","Data":"e2ac402ea256b79cce48c399c07730f98fe1b4de89f6828860449d6f92fb896e"} Dec 08 15:44:07 crc kubenswrapper[4894]: I1208 15:44:07.210837 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58adb9cc-ed4d-4d75-863e-e09680757175" path="/var/lib/kubelet/pods/58adb9cc-ed4d-4d75-863e-e09680757175/volumes" Dec 08 15:44:07 crc kubenswrapper[4894]: I1208 15:44:07.211434 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c4b76/crc-debug-5lxbm"] Dec 08 15:44:07 crc kubenswrapper[4894]: I1208 15:44:07.211468 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c4b76/crc-debug-5lxbm"] Dec 08 15:44:08 crc kubenswrapper[4894]: I1208 15:44:08.266381 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:08 crc kubenswrapper[4894]: I1208 15:44:08.356757 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530a157-fc6c-4207-9725-89f3b2d2806b-host\") pod \"7530a157-fc6c-4207-9725-89f3b2d2806b\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " Dec 08 15:44:08 crc kubenswrapper[4894]: I1208 15:44:08.356865 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29zps\" (UniqueName: \"kubernetes.io/projected/7530a157-fc6c-4207-9725-89f3b2d2806b-kube-api-access-29zps\") pod \"7530a157-fc6c-4207-9725-89f3b2d2806b\" (UID: \"7530a157-fc6c-4207-9725-89f3b2d2806b\") " Dec 08 15:44:08 crc kubenswrapper[4894]: I1208 15:44:08.356952 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7530a157-fc6c-4207-9725-89f3b2d2806b-host" (OuterVolumeSpecName: "host") pod "7530a157-fc6c-4207-9725-89f3b2d2806b" (UID: "7530a157-fc6c-4207-9725-89f3b2d2806b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:44:08 crc kubenswrapper[4894]: I1208 15:44:08.357254 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7530a157-fc6c-4207-9725-89f3b2d2806b-host\") on node \"crc\" DevicePath \"\"" Dec 08 15:44:08 crc kubenswrapper[4894]: I1208 15:44:08.371500 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7530a157-fc6c-4207-9725-89f3b2d2806b-kube-api-access-29zps" (OuterVolumeSpecName: "kube-api-access-29zps") pod "7530a157-fc6c-4207-9725-89f3b2d2806b" (UID: "7530a157-fc6c-4207-9725-89f3b2d2806b"). InnerVolumeSpecName "kube-api-access-29zps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:44:08 crc kubenswrapper[4894]: I1208 15:44:08.458912 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29zps\" (UniqueName: \"kubernetes.io/projected/7530a157-fc6c-4207-9725-89f3b2d2806b-kube-api-access-29zps\") on node \"crc\" DevicePath \"\"" Dec 08 15:44:09 crc kubenswrapper[4894]: I1208 15:44:09.178754 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2ac402ea256b79cce48c399c07730f98fe1b4de89f6828860449d6f92fb896e" Dec 08 15:44:09 crc kubenswrapper[4894]: I1208 15:44:09.178796 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/crc-debug-5lxbm" Dec 08 15:44:09 crc kubenswrapper[4894]: I1208 15:44:09.208303 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7530a157-fc6c-4207-9725-89f3b2d2806b" path="/var/lib/kubelet/pods/7530a157-fc6c-4207-9725-89f3b2d2806b/volumes" Dec 08 15:44:21 crc kubenswrapper[4894]: I1208 15:44:21.926464 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-789cf67ffd-s9792_7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5/barbican-api/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.123199 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-789cf67ffd-s9792_7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5/barbican-api-log/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.154717 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5897b758c4-ntx2t_561cc404-97e2-4808-9c11-e83efcb00484/barbican-keystone-listener-log/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.185327 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5897b758c4-ntx2t_561cc404-97e2-4808-9c11-e83efcb00484/barbican-keystone-listener/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.300478 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-845d6bc787-swrrt_f3b10111-9cb4-4baa-863b-3c34341b2e07/barbican-worker/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.342390 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-845d6bc787-swrrt_f3b10111-9cb4-4baa-863b-3c34341b2e07/barbican-worker-log/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.521655 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph_86e41fca-6958-4b8a-84c0-851b75f8efea/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.605057 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/ceilometer-central-agent/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.678441 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/ceilometer-notification-agent/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.725181 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/proxy-httpd/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.808601 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/sg-core/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.911544 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_11765aa0-3b20-4d49-9017-8a56603c10b0/cinder-api/0.log" Dec 08 15:44:22 crc kubenswrapper[4894]: I1208 15:44:22.918398 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_11765aa0-3b20-4d49-9017-8a56603c10b0/cinder-api-log/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.075478 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f2435a2b-5129-4479-b888-d3874b125874/cinder-scheduler/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.134303 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f2435a2b-5129-4479-b888-d3874b125874/probe/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.251058 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-x999n_e6f24a78-43c1-491e-98d0-3bd4d1d045eb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.359582 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx_0f36f23c-cd37-4fe4-b851-6f1aa2c4017a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.445003 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-nb62g_487bbd0b-1201-4a53-b475-86f85f5a7f51/init/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.636489 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-nb62g_487bbd0b-1201-4a53-b475-86f85f5a7f51/init/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.670169 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z_d2537804-7d9e-41e2-9224-58916974c998/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.699370 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-nb62g_487bbd0b-1201-4a53-b475-86f85f5a7f51/dnsmasq-dns/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.902480 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2d3952ab-f60a-4450-bd40-27b2b99a3e9e/glance-httpd/0.log" Dec 08 15:44:23 crc kubenswrapper[4894]: I1208 15:44:23.930540 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2d3952ab-f60a-4450-bd40-27b2b99a3e9e/glance-log/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.073276 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_62340785-851a-479a-8c14-3908972a9a9f/glance-httpd/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.097522 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_62340785-851a-479a-8c14-3908972a9a9f/glance-log/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.204616 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dcfbfb8-294k4_c3009b99-5e3b-4f39-b62f-0e04aeb371f2/horizon/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.412710 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v_d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.525221 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dcfbfb8-294k4_c3009b99-5e3b-4f39-b62f-0e04aeb371f2/horizon-log/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.605036 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-vcccf_dff601df-c402-489d-b878-a8aaa7a32909/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.846107 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7cb79b8b98-42mc8_57827c7d-7ca8-4eb4-80e9-c6690fca1dac/keystone-api/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.884555 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_5ea146ea-5684-46f7-850c-77df5b66c07b/kube-state-metrics/0.log" Dec 08 15:44:24 crc kubenswrapper[4894]: I1208 15:44:24.986991 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7_12612e53-bc51-4b09-9f07-f00d8debac18/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:25 crc kubenswrapper[4894]: I1208 15:44:25.340433 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5576bf896f-fxztt_70163272-391d-4c33-be21-3e504f9a179f/neutron-api/0.log" Dec 08 15:44:25 crc kubenswrapper[4894]: I1208 15:44:25.361691 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5576bf896f-fxztt_70163272-391d-4c33-be21-3e504f9a179f/neutron-httpd/0.log" Dec 08 15:44:25 crc kubenswrapper[4894]: I1208 15:44:25.590423 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c_35d2beec-8391-420e-a413-44fed1095880/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:26 crc kubenswrapper[4894]: I1208 15:44:26.124072 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_635376cd-70d9-4108-942a-927f52ce27cb/nova-api-api/0.log" Dec 08 15:44:26 crc kubenswrapper[4894]: I1208 15:44:26.215060 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_96255803-4336-4b64-8aee-956302d0c5a2/nova-cell0-conductor-conductor/0.log" Dec 08 15:44:26 crc kubenswrapper[4894]: I1208 15:44:26.267087 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_635376cd-70d9-4108-942a-927f52ce27cb/nova-api-log/0.log" Dec 08 15:44:26 crc kubenswrapper[4894]: I1208 15:44:26.466682 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e939bafe-db68-41d7-8505-651b08bd1ffd/nova-cell1-conductor-conductor/0.log" Dec 08 15:44:26 crc kubenswrapper[4894]: I1208 15:44:26.499425 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1aee5a31-802a-4d05-b242-a7228bc76fbf/nova-cell1-novncproxy-novncproxy/0.log" Dec 08 15:44:26 crc kubenswrapper[4894]: I1208 15:44:26.656113 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ktkvz_01439e20-f7a9-4aa6-bcd7-754e48a7d8b2/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:26 crc kubenswrapper[4894]: I1208 15:44:26.800250 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097/nova-metadata-log/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.097925 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fc76d242-0072-4bef-a299-0a88624cb734/nova-scheduler-scheduler/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.152537 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1e121dc9-bd10-476c-b28b-06c6dcb09165/mysql-bootstrap/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.332933 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1e121dc9-bd10-476c-b28b-06c6dcb09165/galera/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.340463 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1e121dc9-bd10-476c-b28b-06c6dcb09165/mysql-bootstrap/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.499629 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_755a6729-c028-4206-abc7-5b4bd3825a3e/mysql-bootstrap/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.731416 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_755a6729-c028-4206-abc7-5b4bd3825a3e/mysql-bootstrap/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.770931 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_755a6729-c028-4206-abc7-5b4bd3825a3e/galera/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.889221 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097/nova-metadata-metadata/0.log" Dec 08 15:44:27 crc kubenswrapper[4894]: I1208 15:44:27.937324 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_29e0f64a-aac1-4909-b7f6-d83223c04d31/openstackclient/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.014088 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-89p8p_9d8ef054-d4aa-482b-abad-c961a2a5f2ad/ovn-controller/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.162217 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-j22jr_2150dea7-8071-418c-ba2d-20eeb73f2ff8/openstack-network-exporter/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.225449 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovsdb-server-init/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.428587 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovsdb-server-init/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.441864 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovsdb-server/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.496857 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovs-vswitchd/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.719805 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_65f9ac8b-3288-4ea6-abc7-c396fb830f69/openstack-network-exporter/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.728041 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-fvbzv_0ccc212f-4efb-4444-9156-27751b81562e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.813313 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_65f9ac8b-3288-4ea6-abc7-c396fb830f69/ovn-northd/0.log" Dec 08 15:44:28 crc kubenswrapper[4894]: I1208 15:44:28.935676 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_0949aa02-5e23-4bd9-9a93-180a59f701ad/openstack-network-exporter/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.013359 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_0949aa02-5e23-4bd9-9a93-180a59f701ad/ovsdbserver-nb/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.134796 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_31bac65c-93e6-4816-837d-a0f467bb3966/openstack-network-exporter/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.193142 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_31bac65c-93e6-4816-837d-a0f467bb3966/ovsdbserver-sb/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.563691 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f74bf4b48-xdrm8_cabb5ceb-58b0-47b5-8db0-cb717c6d8aad/placement-api/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.573509 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ba215854-1f7f-4315-acf0-edd51b2358cb/setup-container/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.579764 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f74bf4b48-xdrm8_cabb5ceb-58b0-47b5-8db0-cb717c6d8aad/placement-log/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.757288 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ba215854-1f7f-4315-acf0-edd51b2358cb/rabbitmq/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.764496 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ba215854-1f7f-4315-acf0-edd51b2358cb/setup-container/0.log" Dec 08 15:44:29 crc kubenswrapper[4894]: I1208 15:44:29.813705 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8447e7e6-4a92-432c-8a5f-48a4cd10675b/setup-container/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.042756 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8447e7e6-4a92-432c-8a5f-48a4cd10675b/setup-container/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.043615 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8447e7e6-4a92-432c-8a5f-48a4cd10675b/rabbitmq/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.070043 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv_fae74a7d-767f-47a9-8f3c-1337c1be1d3e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.268119 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qsrmq_0147e712-181c-4b8e-85d3-3b90c985f5ee/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.405648 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7_211799eb-bf07-4965-8404-ec3ddba08c3b/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.524648 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lxvp6_f840f001-3515-40c7-bc6e-6434ee8df42a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.638096 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-m87rg_b5b86fe7-04aa-4538-88d1-d6dd42673b07/ssh-known-hosts-edpm-deployment/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.761342 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-567d6f4bf7-xrnjd_c7a0f531-0d8b-4858-aee9-981bdfbc81e0/proxy-server/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.843194 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-567d6f4bf7-xrnjd_c7a0f531-0d8b-4858-aee9-981bdfbc81e0/proxy-httpd/0.log" Dec 08 15:44:30 crc kubenswrapper[4894]: I1208 15:44:30.984203 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m4f8h_daf51f62-aa82-474b-b3d4-4746043cf399/swift-ring-rebalance/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.034169 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-auditor/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.117797 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-reaper/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.197384 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-replicator/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.265479 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-server/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.281792 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-auditor/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.387621 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-replicator/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.429421 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-server/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.459037 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-updater/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.538247 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-auditor/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.591425 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-expirer/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.644524 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-replicator/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.690166 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-server/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.734357 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-updater/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.844451 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/rsync/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.851309 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/swift-recon-cron/0.log" Dec 08 15:44:31 crc kubenswrapper[4894]: I1208 15:44:31.967354 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp_1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:32 crc kubenswrapper[4894]: I1208 15:44:32.074252 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_39de6187-9721-4687-8001-ef13c4abafb3/tempest-tests-tempest-tests-runner/0.log" Dec 08 15:44:32 crc kubenswrapper[4894]: I1208 15:44:32.228547 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb/test-operator-logs-container/0.log" Dec 08 15:44:32 crc kubenswrapper[4894]: I1208 15:44:32.312916 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj_18936dba-63ba-44de-954c-c65df5b5743a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:44:39 crc kubenswrapper[4894]: I1208 15:44:39.798059 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_637635c4-70df-45dd-8139-9c8dc3f53ea0/memcached/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.068095 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-kh5v9_28cfecb9-9c7c-49ae-af33-fc2d9acd374a/kube-rbac-proxy/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.124331 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-kh5v9_28cfecb9-9c7c-49ae-af33-fc2d9acd374a/manager/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.291048 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8vb7c_4ee4db87-8c97-4002-844b-6c55e1a41b2e/kube-rbac-proxy/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.318486 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8vb7c_4ee4db87-8c97-4002-844b-6c55e1a41b2e/manager/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.424022 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-f67zc_e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72/kube-rbac-proxy/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.489719 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-f67zc_e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72/manager/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.562365 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/util/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.724855 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/pull/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.742218 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/pull/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.748305 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/util/0.log" Dec 08 15:44:56 crc kubenswrapper[4894]: I1208 15:44:56.993874 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/util/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.097451 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/extract/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.111464 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/pull/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.227714 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-ffvpw_b8a908e6-3df2-4dc9-9c9c-068e91dc839c/kube-rbac-proxy/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.343512 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-ffvpw_b8a908e6-3df2-4dc9-9c9c-068e91dc839c/manager/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.349931 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-frz8t_5d7b9bb8-c977-45cb-b198-5f047059a0a4/kube-rbac-proxy/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.518997 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-frz8t_5d7b9bb8-c977-45cb-b198-5f047059a0a4/manager/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.524302 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-mtkrb_afda731d-73a8-443f-9647-a774ca8c3146/kube-rbac-proxy/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.570233 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-mtkrb_afda731d-73a8-443f-9647-a774ca8c3146/manager/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.710462 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4xrts_c3e89da4-2a46-4982-9392-b0cf87a1edc3/kube-rbac-proxy/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.917339 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-47bpp_f7ec7a43-f5b5-474f-af15-d0e664699477/kube-rbac-proxy/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.934844 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4xrts_c3e89da4-2a46-4982-9392-b0cf87a1edc3/manager/0.log" Dec 08 15:44:57 crc kubenswrapper[4894]: I1208 15:44:57.970115 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-47bpp_f7ec7a43-f5b5-474f-af15-d0e664699477/manager/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.113992 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-lbszt_2df93571-3b4e-4ab0-ad8b-34d2ef76034c/kube-rbac-proxy/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.186950 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-lbszt_2df93571-3b4e-4ab0-ad8b-34d2ef76034c/manager/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.284579 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w7xbj_aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2/kube-rbac-proxy/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.292232 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w7xbj_aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2/manager/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.389015 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-h2xw4_e6585a6c-b50d-41d3-a080-ab7cb27f9199/kube-rbac-proxy/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.477004 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-h2xw4_e6585a6c-b50d-41d3-a080-ab7cb27f9199/manager/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.563575 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2sfqd_db804bba-5a84-4d23-93ca-e7118afc2af8/kube-rbac-proxy/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.637547 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2sfqd_db804bba-5a84-4d23-93ca-e7118afc2af8/manager/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.724415 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jc2rb_c91237dd-8bb9-4240-b24a-e19b58687d2e/kube-rbac-proxy/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.847247 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jc2rb_c91237dd-8bb9-4240-b24a-e19b58687d2e/manager/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.908857 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cqq4p_750c567a-693a-401b-8e7f-d8f1befbeacb/kube-rbac-proxy/0.log" Dec 08 15:44:58 crc kubenswrapper[4894]: I1208 15:44:58.985353 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cqq4p_750c567a-693a-401b-8e7f-d8f1befbeacb/manager/0.log" Dec 08 15:44:59 crc kubenswrapper[4894]: I1208 15:44:59.088073 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4lcw4_fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae/kube-rbac-proxy/0.log" Dec 08 15:44:59 crc kubenswrapper[4894]: I1208 15:44:59.136174 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4lcw4_fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae/manager/0.log" Dec 08 15:44:59 crc kubenswrapper[4894]: I1208 15:44:59.501182 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8ptfg_654a1e08-5902-4a0f-8584-f662bc005210/registry-server/0.log" Dec 08 15:44:59 crc kubenswrapper[4894]: I1208 15:44:59.549971 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-647974bbd-56w9g_d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd/operator/0.log" Dec 08 15:44:59 crc kubenswrapper[4894]: I1208 15:44:59.682397 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lr526_2b5f3682-eb12-440c-a016-f165ec4e7cfa/kube-rbac-proxy/0.log" Dec 08 15:44:59 crc kubenswrapper[4894]: I1208 15:44:59.842363 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lr526_2b5f3682-eb12-440c-a016-f165ec4e7cfa/manager/0.log" Dec 08 15:44:59 crc kubenswrapper[4894]: I1208 15:44:59.962797 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-sxfkp_59051068-3252-47e6-a826-79d1951c3d5c/kube-rbac-proxy/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.075138 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-sxfkp_59051068-3252-47e6-a826-79d1951c3d5c/manager/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.155734 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8"] Dec 08 15:45:00 crc kubenswrapper[4894]: E1208 15:45:00.156187 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7530a157-fc6c-4207-9725-89f3b2d2806b" containerName="container-00" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.156200 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7530a157-fc6c-4207-9725-89f3b2d2806b" containerName="container-00" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.156421 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7530a157-fc6c-4207-9725-89f3b2d2806b" containerName="container-00" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.157076 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.162215 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.167727 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.173884 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlscl\" (UniqueName: \"kubernetes.io/projected/e725abbf-fcad-4e55-af88-e2e8f05eaf00-kube-api-access-vlscl\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.174022 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e725abbf-fcad-4e55-af88-e2e8f05eaf00-secret-volume\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.174155 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e725abbf-fcad-4e55-af88-e2e8f05eaf00-config-volume\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.190094 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8"] Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.275770 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlscl\" (UniqueName: \"kubernetes.io/projected/e725abbf-fcad-4e55-af88-e2e8f05eaf00-kube-api-access-vlscl\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.275887 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e725abbf-fcad-4e55-af88-e2e8f05eaf00-secret-volume\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.275972 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e725abbf-fcad-4e55-af88-e2e8f05eaf00-config-volume\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.279090 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e725abbf-fcad-4e55-af88-e2e8f05eaf00-config-volume\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.283255 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-97mkf_9c6f4113-6d2f-4fce-b76b-8802b3b605c6/operator/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.294369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e725abbf-fcad-4e55-af88-e2e8f05eaf00-secret-volume\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.299486 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlscl\" (UniqueName: \"kubernetes.io/projected/e725abbf-fcad-4e55-af88-e2e8f05eaf00-kube-api-access-vlscl\") pod \"collect-profiles-29420145-ltjc8\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.344149 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-69wkl_5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8/kube-rbac-proxy/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.421955 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-69wkl_5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8/manager/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.446619 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8695b4b66-c64t7_69d7f3da-48c5-47e8-853b-0cca6246d96e/manager/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.503844 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.531832 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-6qvtl_2536d7d4-e60f-44a6-8314-cfd34b5545e9/kube-rbac-proxy/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.637785 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-6qvtl_2536d7d4-e60f-44a6-8314-cfd34b5545e9/manager/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.757750 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-zhwdn_8b8eda01-4bc4-4eca-b909-901024ef466b/kube-rbac-proxy/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.819349 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-zhwdn_8b8eda01-4bc4-4eca-b909-901024ef466b/manager/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.907442 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-rnqzt_8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76/kube-rbac-proxy/0.log" Dec 08 15:45:00 crc kubenswrapper[4894]: I1208 15:45:00.908565 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-rnqzt_8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76/manager/0.log" Dec 08 15:45:01 crc kubenswrapper[4894]: I1208 15:45:01.086229 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8"] Dec 08 15:45:01 crc kubenswrapper[4894]: I1208 15:45:01.629406 4894 generic.go:334] "Generic (PLEG): container finished" podID="e725abbf-fcad-4e55-af88-e2e8f05eaf00" containerID="60ab4e0fddfc8fe0a351893e36be3a1a1c859dd33b641ca5a618e34698a0b7fe" exitCode=0 Dec 08 15:45:01 crc kubenswrapper[4894]: I1208 15:45:01.629453 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" event={"ID":"e725abbf-fcad-4e55-af88-e2e8f05eaf00","Type":"ContainerDied","Data":"60ab4e0fddfc8fe0a351893e36be3a1a1c859dd33b641ca5a618e34698a0b7fe"} Dec 08 15:45:01 crc kubenswrapper[4894]: I1208 15:45:01.629479 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" event={"ID":"e725abbf-fcad-4e55-af88-e2e8f05eaf00","Type":"ContainerStarted","Data":"0d6183d7c514b54ff388d6d16186f1c4abce18db4655642c8d74d1e54350f97f"} Dec 08 15:45:02 crc kubenswrapper[4894]: I1208 15:45:02.957252 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.039733 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlscl\" (UniqueName: \"kubernetes.io/projected/e725abbf-fcad-4e55-af88-e2e8f05eaf00-kube-api-access-vlscl\") pod \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.039867 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e725abbf-fcad-4e55-af88-e2e8f05eaf00-secret-volume\") pod \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.039893 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e725abbf-fcad-4e55-af88-e2e8f05eaf00-config-volume\") pod \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\" (UID: \"e725abbf-fcad-4e55-af88-e2e8f05eaf00\") " Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.041036 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e725abbf-fcad-4e55-af88-e2e8f05eaf00-config-volume" (OuterVolumeSpecName: "config-volume") pod "e725abbf-fcad-4e55-af88-e2e8f05eaf00" (UID: "e725abbf-fcad-4e55-af88-e2e8f05eaf00"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.048151 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e725abbf-fcad-4e55-af88-e2e8f05eaf00-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e725abbf-fcad-4e55-af88-e2e8f05eaf00" (UID: "e725abbf-fcad-4e55-af88-e2e8f05eaf00"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.048740 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e725abbf-fcad-4e55-af88-e2e8f05eaf00-kube-api-access-vlscl" (OuterVolumeSpecName: "kube-api-access-vlscl") pod "e725abbf-fcad-4e55-af88-e2e8f05eaf00" (UID: "e725abbf-fcad-4e55-af88-e2e8f05eaf00"). InnerVolumeSpecName "kube-api-access-vlscl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.141458 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e725abbf-fcad-4e55-af88-e2e8f05eaf00-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.141493 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e725abbf-fcad-4e55-af88-e2e8f05eaf00-config-volume\") on node \"crc\" DevicePath \"\"" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.141503 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlscl\" (UniqueName: \"kubernetes.io/projected/e725abbf-fcad-4e55-af88-e2e8f05eaf00-kube-api-access-vlscl\") on node \"crc\" DevicePath \"\"" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.647306 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" event={"ID":"e725abbf-fcad-4e55-af88-e2e8f05eaf00","Type":"ContainerDied","Data":"0d6183d7c514b54ff388d6d16186f1c4abce18db4655642c8d74d1e54350f97f"} Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.647360 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d6183d7c514b54ff388d6d16186f1c4abce18db4655642c8d74d1e54350f97f" Dec 08 15:45:03 crc kubenswrapper[4894]: I1208 15:45:03.647663 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29420145-ltjc8" Dec 08 15:45:04 crc kubenswrapper[4894]: I1208 15:45:04.029759 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk"] Dec 08 15:45:04 crc kubenswrapper[4894]: I1208 15:45:04.040671 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29420100-pgpxk"] Dec 08 15:45:05 crc kubenswrapper[4894]: I1208 15:45:05.209630 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f324fd-5585-43ac-9045-cc94aeb37fac" path="/var/lib/kubelet/pods/b9f324fd-5585-43ac-9045-cc94aeb37fac/volumes" Dec 08 15:45:19 crc kubenswrapper[4894]: I1208 15:45:19.648993 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-njrj9_e899bcbb-0ffd-4989-bba4-47b61fe07832/control-plane-machine-set-operator/0.log" Dec 08 15:45:19 crc kubenswrapper[4894]: I1208 15:45:19.896633 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7nbxt_22a82f00-b781-4068-b398-0b4c738db413/machine-api-operator/0.log" Dec 08 15:45:19 crc kubenswrapper[4894]: I1208 15:45:19.904962 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7nbxt_22a82f00-b781-4068-b398-0b4c738db413/kube-rbac-proxy/0.log" Dec 08 15:45:32 crc kubenswrapper[4894]: I1208 15:45:32.105304 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nzrw5_0aa35d3c-6f53-415d-8a11-f2e8cc6298fb/cert-manager-controller/0.log" Dec 08 15:45:32 crc kubenswrapper[4894]: I1208 15:45:32.264893 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qfw89_960dc174-5cd7-4931-94f7-79871062fcb9/cert-manager-cainjector/0.log" Dec 08 15:45:32 crc kubenswrapper[4894]: I1208 15:45:32.326661 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hzvvg_7774fcfd-7cfb-4b62-8e29-48c16c5d0337/cert-manager-webhook/0.log" Dec 08 15:45:37 crc kubenswrapper[4894]: I1208 15:45:37.296558 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:45:37 crc kubenswrapper[4894]: I1208 15:45:37.297949 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:45:44 crc kubenswrapper[4894]: I1208 15:45:44.164760 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5v49c_55c7d1cf-5cb4-4a06-ac50-2ee06f72a522/nmstate-console-plugin/0.log" Dec 08 15:45:44 crc kubenswrapper[4894]: I1208 15:45:44.326576 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-b8g67_2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c/nmstate-handler/0.log" Dec 08 15:45:44 crc kubenswrapper[4894]: I1208 15:45:44.402295 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lkw5q_2f8534ee-633a-4633-9649-dbc49a705529/nmstate-metrics/0.log" Dec 08 15:45:44 crc kubenswrapper[4894]: I1208 15:45:44.411790 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lkw5q_2f8534ee-633a-4633-9649-dbc49a705529/kube-rbac-proxy/0.log" Dec 08 15:45:44 crc kubenswrapper[4894]: I1208 15:45:44.612428 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-7kjmb_68045b7d-a426-4739-9870-b4e40c54dedd/nmstate-operator/0.log" Dec 08 15:45:44 crc kubenswrapper[4894]: I1208 15:45:44.664150 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-9wlhf_70fa62b1-a4b9-49c3-abd2-49ac96c501b6/nmstate-webhook/0.log" Dec 08 15:45:58 crc kubenswrapper[4894]: I1208 15:45:58.288217 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6m7rl_e2473db3-776f-4a3d-9b4f-aba30d96463b/kube-rbac-proxy/0.log" Dec 08 15:45:58 crc kubenswrapper[4894]: I1208 15:45:58.404279 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6m7rl_e2473db3-776f-4a3d-9b4f-aba30d96463b/controller/0.log" Dec 08 15:45:58 crc kubenswrapper[4894]: I1208 15:45:58.679713 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:45:58 crc kubenswrapper[4894]: I1208 15:45:58.831497 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:45:58 crc kubenswrapper[4894]: I1208 15:45:58.852418 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:45:58 crc kubenswrapper[4894]: I1208 15:45:58.869020 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:45:58 crc kubenswrapper[4894]: I1208 15:45:58.919180 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.077856 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.097959 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.111997 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.115138 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.182416 4894 scope.go:117] "RemoveContainer" containerID="966763b2cd33c4d7db822327b65c6783eeed176e05227011693e39bf8f6abcfa" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.336621 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.343662 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.373327 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.386266 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/controller/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.522373 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/frr-metrics/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.573248 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/kube-rbac-proxy-frr/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.574946 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/kube-rbac-proxy/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.742228 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/reloader/0.log" Dec 08 15:45:59 crc kubenswrapper[4894]: I1208 15:45:59.795592 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-rt5h4_29388ab6-9e49-4662-8d74-1dda6e8d228e/frr-k8s-webhook-server/0.log" Dec 08 15:46:00 crc kubenswrapper[4894]: I1208 15:46:00.045199 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-78884d4998-whjq5_10bcfd29-cff4-4fcc-be38-43296e67cac7/manager/0.log" Dec 08 15:46:00 crc kubenswrapper[4894]: I1208 15:46:00.165693 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74cdf75695-th2w6_6e5ec2a2-054b-4cb6-a3d6-7376f254869c/webhook-server/0.log" Dec 08 15:46:00 crc kubenswrapper[4894]: I1208 15:46:00.280576 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hxwz6_e316690b-6982-4588-ab38-a33ca86ba1f0/kube-rbac-proxy/0.log" Dec 08 15:46:00 crc kubenswrapper[4894]: I1208 15:46:00.882829 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/frr/0.log" Dec 08 15:46:00 crc kubenswrapper[4894]: I1208 15:46:00.889023 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hxwz6_e316690b-6982-4588-ab38-a33ca86ba1f0/speaker/0.log" Dec 08 15:46:07 crc kubenswrapper[4894]: I1208 15:46:07.297394 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:46:07 crc kubenswrapper[4894]: I1208 15:46:07.297837 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:46:12 crc kubenswrapper[4894]: I1208 15:46:12.871321 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/util/0.log" Dec 08 15:46:12 crc kubenswrapper[4894]: I1208 15:46:12.945574 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/util/0.log" Dec 08 15:46:12 crc kubenswrapper[4894]: I1208 15:46:12.992020 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/pull/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.078891 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/pull/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.218291 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/util/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.240180 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/pull/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.249747 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/extract/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.366728 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/util/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.547843 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/pull/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.549919 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/pull/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.577953 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/util/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.738085 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/pull/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.742869 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/util/0.log" Dec 08 15:46:13 crc kubenswrapper[4894]: I1208 15:46:13.747643 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/extract/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.092548 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-utilities/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.222284 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-content/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.224308 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-utilities/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.240269 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-content/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.396299 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-utilities/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.424536 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-content/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.635003 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7cr5_9393c174-85f4-4267-80a1-e7c1075104b5/extract-utilities/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.766541 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7cr5_9393c174-85f4-4267-80a1-e7c1075104b5/extract-utilities/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.826283 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7cr5_9393c174-85f4-4267-80a1-e7c1075104b5/extract-content/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.852057 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/registry-server/0.log" Dec 08 15:46:14 crc kubenswrapper[4894]: I1208 15:46:14.863589 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7cr5_9393c174-85f4-4267-80a1-e7c1075104b5/extract-content/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.096986 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7cr5_9393c174-85f4-4267-80a1-e7c1075104b5/extract-utilities/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.097152 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7cr5_9393c174-85f4-4267-80a1-e7c1075104b5/extract-content/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.338698 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bl7db_e1fa33df-7fab-40bc-8a30-9004002f770a/marketplace-operator/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.385562 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-utilities/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.704324 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s7cr5_9393c174-85f4-4267-80a1-e7c1075104b5/registry-server/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.721026 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-content/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.729287 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-utilities/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.771008 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-content/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.925203 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-utilities/0.log" Dec 08 15:46:15 crc kubenswrapper[4894]: I1208 15:46:15.932335 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-content/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.072080 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/registry-server/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.161440 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-utilities/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.340657 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-content/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.368307 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-content/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.377725 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-utilities/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.534924 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-utilities/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.543097 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-content/0.log" Dec 08 15:46:16 crc kubenswrapper[4894]: I1208 15:46:16.788082 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/registry-server/0.log" Dec 08 15:46:36 crc kubenswrapper[4894]: E1208 15:46:36.014921 4894 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.51:35820->38.102.83.51:44119: read tcp 38.102.83.51:35820->38.102.83.51:44119: read: connection reset by peer Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.296692 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.297035 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.297232 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.298092 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.298162 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" gracePeriod=600 Dec 08 15:46:37 crc kubenswrapper[4894]: E1208 15:46:37.425315 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.480461 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" exitCode=0 Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.480503 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295"} Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.480590 4894 scope.go:117] "RemoveContainer" containerID="65921f34e15335690428cdddc29012738b164fb0aea2b952bd02813b0a2dad6e" Dec 08 15:46:37 crc kubenswrapper[4894]: I1208 15:46:37.481207 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:46:37 crc kubenswrapper[4894]: E1208 15:46:37.481576 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:46:47 crc kubenswrapper[4894]: E1208 15:46:47.833093 4894 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.51:53654->38.102.83.51:44119: read tcp 38.102.83.51:53654->38.102.83.51:44119: read: connection reset by peer Dec 08 15:46:51 crc kubenswrapper[4894]: I1208 15:46:51.197876 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:46:51 crc kubenswrapper[4894]: E1208 15:46:51.199069 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:47:03 crc kubenswrapper[4894]: I1208 15:47:03.197059 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:47:03 crc kubenswrapper[4894]: E1208 15:47:03.197828 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.196857 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jb4wg"] Dec 08 15:47:14 crc kubenswrapper[4894]: E1208 15:47:14.197963 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e725abbf-fcad-4e55-af88-e2e8f05eaf00" containerName="collect-profiles" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.197980 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e725abbf-fcad-4e55-af88-e2e8f05eaf00" containerName="collect-profiles" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.198208 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e725abbf-fcad-4e55-af88-e2e8f05eaf00" containerName="collect-profiles" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.200283 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.229485 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jb4wg"] Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.296309 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mhv4\" (UniqueName: \"kubernetes.io/projected/458c93d2-67fd-4afe-9caf-6364d702d0ad-kube-api-access-2mhv4\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.296569 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-utilities\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.296779 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-catalog-content\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.398357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-catalog-content\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.398780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mhv4\" (UniqueName: \"kubernetes.io/projected/458c93d2-67fd-4afe-9caf-6364d702d0ad-kube-api-access-2mhv4\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.399028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-utilities\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.399095 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-catalog-content\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.399473 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-utilities\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.417727 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mhv4\" (UniqueName: \"kubernetes.io/projected/458c93d2-67fd-4afe-9caf-6364d702d0ad-kube-api-access-2mhv4\") pod \"redhat-operators-jb4wg\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:14 crc kubenswrapper[4894]: I1208 15:47:14.525131 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:15 crc kubenswrapper[4894]: I1208 15:47:15.019358 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jb4wg"] Dec 08 15:47:15 crc kubenswrapper[4894]: I1208 15:47:15.878526 4894 generic.go:334] "Generic (PLEG): container finished" podID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerID="9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77" exitCode=0 Dec 08 15:47:15 crc kubenswrapper[4894]: I1208 15:47:15.878636 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jb4wg" event={"ID":"458c93d2-67fd-4afe-9caf-6364d702d0ad","Type":"ContainerDied","Data":"9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77"} Dec 08 15:47:15 crc kubenswrapper[4894]: I1208 15:47:15.878895 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jb4wg" event={"ID":"458c93d2-67fd-4afe-9caf-6364d702d0ad","Type":"ContainerStarted","Data":"907f1ad6feed36802cfbdac0ab6933256b2b9c1906f01415e7bdb808f58b0769"} Dec 08 15:47:16 crc kubenswrapper[4894]: I1208 15:47:16.890497 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jb4wg" event={"ID":"458c93d2-67fd-4afe-9caf-6364d702d0ad","Type":"ContainerStarted","Data":"72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7"} Dec 08 15:47:17 crc kubenswrapper[4894]: I1208 15:47:17.923183 4894 generic.go:334] "Generic (PLEG): container finished" podID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerID="72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7" exitCode=0 Dec 08 15:47:17 crc kubenswrapper[4894]: I1208 15:47:17.923278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jb4wg" event={"ID":"458c93d2-67fd-4afe-9caf-6364d702d0ad","Type":"ContainerDied","Data":"72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7"} Dec 08 15:47:18 crc kubenswrapper[4894]: I1208 15:47:18.197846 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:47:18 crc kubenswrapper[4894]: E1208 15:47:18.198245 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:47:18 crc kubenswrapper[4894]: I1208 15:47:18.934472 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jb4wg" event={"ID":"458c93d2-67fd-4afe-9caf-6364d702d0ad","Type":"ContainerStarted","Data":"d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171"} Dec 08 15:47:18 crc kubenswrapper[4894]: I1208 15:47:18.959953 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jb4wg" podStartSLOduration=2.51779602 podStartE2EDuration="4.959936888s" podCreationTimestamp="2025-12-08 15:47:14 +0000 UTC" firstStartedPulling="2025-12-08 15:47:15.88038499 +0000 UTC m=+3656.980391105" lastFinishedPulling="2025-12-08 15:47:18.322525858 +0000 UTC m=+3659.422531973" observedRunningTime="2025-12-08 15:47:18.949396058 +0000 UTC m=+3660.049402173" watchObservedRunningTime="2025-12-08 15:47:18.959936888 +0000 UTC m=+3660.059943003" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.555291 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h7mv4"] Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.558501 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.567771 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h7mv4"] Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.668447 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-catalog-content\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.668520 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-utilities\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.668669 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mbw7\" (UniqueName: \"kubernetes.io/projected/59da35a4-4f0e-4a94-b90d-ff95b13ed602-kube-api-access-6mbw7\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.770724 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-utilities\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.770875 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mbw7\" (UniqueName: \"kubernetes.io/projected/59da35a4-4f0e-4a94-b90d-ff95b13ed602-kube-api-access-6mbw7\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.771051 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-catalog-content\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.771545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-utilities\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.771580 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-catalog-content\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.793178 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mbw7\" (UniqueName: \"kubernetes.io/projected/59da35a4-4f0e-4a94-b90d-ff95b13ed602-kube-api-access-6mbw7\") pod \"redhat-marketplace-h7mv4\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:21 crc kubenswrapper[4894]: I1208 15:47:21.890849 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:22 crc kubenswrapper[4894]: I1208 15:47:22.201354 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h7mv4"] Dec 08 15:47:22 crc kubenswrapper[4894]: I1208 15:47:22.967658 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerStarted","Data":"1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629"} Dec 08 15:47:22 crc kubenswrapper[4894]: I1208 15:47:22.967706 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerStarted","Data":"77794e09b7fa7bbcae7711a66ae4fff6a0c77bdf00d9b6cd24d7e9391a328ea2"} Dec 08 15:47:23 crc kubenswrapper[4894]: I1208 15:47:23.979328 4894 generic.go:334] "Generic (PLEG): container finished" podID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerID="1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629" exitCode=0 Dec 08 15:47:23 crc kubenswrapper[4894]: I1208 15:47:23.979401 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerDied","Data":"1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629"} Dec 08 15:47:24 crc kubenswrapper[4894]: I1208 15:47:24.525718 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:24 crc kubenswrapper[4894]: I1208 15:47:24.525763 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:24 crc kubenswrapper[4894]: I1208 15:47:24.587430 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:24 crc kubenswrapper[4894]: I1208 15:47:24.992516 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerStarted","Data":"483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36"} Dec 08 15:47:25 crc kubenswrapper[4894]: I1208 15:47:25.044046 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:26 crc kubenswrapper[4894]: I1208 15:47:26.004478 4894 generic.go:334] "Generic (PLEG): container finished" podID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerID="483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36" exitCode=0 Dec 08 15:47:26 crc kubenswrapper[4894]: I1208 15:47:26.004532 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerDied","Data":"483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36"} Dec 08 15:47:26 crc kubenswrapper[4894]: I1208 15:47:26.945649 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jb4wg"] Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.015782 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerStarted","Data":"7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11"} Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.016798 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jb4wg" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="registry-server" containerID="cri-o://d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171" gracePeriod=2 Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.051891 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h7mv4" podStartSLOduration=3.638514002 podStartE2EDuration="6.051868812s" podCreationTimestamp="2025-12-08 15:47:21 +0000 UTC" firstStartedPulling="2025-12-08 15:47:23.981979727 +0000 UTC m=+3665.081985852" lastFinishedPulling="2025-12-08 15:47:26.395334547 +0000 UTC m=+3667.495340662" observedRunningTime="2025-12-08 15:47:27.042293434 +0000 UTC m=+3668.142299549" watchObservedRunningTime="2025-12-08 15:47:27.051868812 +0000 UTC m=+3668.151874917" Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.478082 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.614541 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mhv4\" (UniqueName: \"kubernetes.io/projected/458c93d2-67fd-4afe-9caf-6364d702d0ad-kube-api-access-2mhv4\") pod \"458c93d2-67fd-4afe-9caf-6364d702d0ad\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.614750 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-catalog-content\") pod \"458c93d2-67fd-4afe-9caf-6364d702d0ad\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.614776 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-utilities\") pod \"458c93d2-67fd-4afe-9caf-6364d702d0ad\" (UID: \"458c93d2-67fd-4afe-9caf-6364d702d0ad\") " Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.615588 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-utilities" (OuterVolumeSpecName: "utilities") pod "458c93d2-67fd-4afe-9caf-6364d702d0ad" (UID: "458c93d2-67fd-4afe-9caf-6364d702d0ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.620042 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458c93d2-67fd-4afe-9caf-6364d702d0ad-kube-api-access-2mhv4" (OuterVolumeSpecName: "kube-api-access-2mhv4") pod "458c93d2-67fd-4afe-9caf-6364d702d0ad" (UID: "458c93d2-67fd-4afe-9caf-6364d702d0ad"). InnerVolumeSpecName "kube-api-access-2mhv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.717101 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:47:27 crc kubenswrapper[4894]: I1208 15:47:27.717306 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mhv4\" (UniqueName: \"kubernetes.io/projected/458c93d2-67fd-4afe-9caf-6364d702d0ad-kube-api-access-2mhv4\") on node \"crc\" DevicePath \"\"" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.027935 4894 generic.go:334] "Generic (PLEG): container finished" podID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerID="d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171" exitCode=0 Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.028109 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jb4wg" event={"ID":"458c93d2-67fd-4afe-9caf-6364d702d0ad","Type":"ContainerDied","Data":"d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171"} Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.028184 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jb4wg" event={"ID":"458c93d2-67fd-4afe-9caf-6364d702d0ad","Type":"ContainerDied","Data":"907f1ad6feed36802cfbdac0ab6933256b2b9c1906f01415e7bdb808f58b0769"} Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.028218 4894 scope.go:117] "RemoveContainer" containerID="d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.028886 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jb4wg" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.066263 4894 scope.go:117] "RemoveContainer" containerID="72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.091690 4894 scope.go:117] "RemoveContainer" containerID="9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.137732 4894 scope.go:117] "RemoveContainer" containerID="d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171" Dec 08 15:47:28 crc kubenswrapper[4894]: E1208 15:47:28.138259 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171\": container with ID starting with d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171 not found: ID does not exist" containerID="d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.138307 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171"} err="failed to get container status \"d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171\": rpc error: code = NotFound desc = could not find container \"d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171\": container with ID starting with d88537b290b0fe5dd37d6215e0e4088b0768622e73c0d240acc5b0830bdf8171 not found: ID does not exist" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.138337 4894 scope.go:117] "RemoveContainer" containerID="72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7" Dec 08 15:47:28 crc kubenswrapper[4894]: E1208 15:47:28.139061 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7\": container with ID starting with 72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7 not found: ID does not exist" containerID="72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.139105 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7"} err="failed to get container status \"72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7\": rpc error: code = NotFound desc = could not find container \"72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7\": container with ID starting with 72c7f42df55489d547b1a19c0532b700aa5fc1a73be0b124a7d92d7062a621e7 not found: ID does not exist" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.139132 4894 scope.go:117] "RemoveContainer" containerID="9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77" Dec 08 15:47:28 crc kubenswrapper[4894]: E1208 15:47:28.139457 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77\": container with ID starting with 9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77 not found: ID does not exist" containerID="9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.139505 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77"} err="failed to get container status \"9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77\": rpc error: code = NotFound desc = could not find container \"9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77\": container with ID starting with 9dfa1789989b10305f9bdb459af9c5ede3c9a064cac18b4ed7785d96601bea77 not found: ID does not exist" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.716790 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "458c93d2-67fd-4afe-9caf-6364d702d0ad" (UID: "458c93d2-67fd-4afe-9caf-6364d702d0ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.737784 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/458c93d2-67fd-4afe-9caf-6364d702d0ad-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.967205 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jb4wg"] Dec 08 15:47:28 crc kubenswrapper[4894]: I1208 15:47:28.974215 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jb4wg"] Dec 08 15:47:29 crc kubenswrapper[4894]: I1208 15:47:29.207960 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" path="/var/lib/kubelet/pods/458c93d2-67fd-4afe-9caf-6364d702d0ad/volumes" Dec 08 15:47:31 crc kubenswrapper[4894]: I1208 15:47:31.197353 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:47:31 crc kubenswrapper[4894]: E1208 15:47:31.197872 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:47:31 crc kubenswrapper[4894]: I1208 15:47:31.892342 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:31 crc kubenswrapper[4894]: I1208 15:47:31.894020 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:31 crc kubenswrapper[4894]: I1208 15:47:31.954733 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:32 crc kubenswrapper[4894]: I1208 15:47:32.114376 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:32 crc kubenswrapper[4894]: I1208 15:47:32.953360 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h7mv4"] Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.097775 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h7mv4" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="registry-server" containerID="cri-o://7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11" gracePeriod=2 Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.689367 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.768431 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-utilities\") pod \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.768506 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-catalog-content\") pod \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.768875 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mbw7\" (UniqueName: \"kubernetes.io/projected/59da35a4-4f0e-4a94-b90d-ff95b13ed602-kube-api-access-6mbw7\") pod \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\" (UID: \"59da35a4-4f0e-4a94-b90d-ff95b13ed602\") " Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.769311 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-utilities" (OuterVolumeSpecName: "utilities") pod "59da35a4-4f0e-4a94-b90d-ff95b13ed602" (UID: "59da35a4-4f0e-4a94-b90d-ff95b13ed602"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.776546 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59da35a4-4f0e-4a94-b90d-ff95b13ed602-kube-api-access-6mbw7" (OuterVolumeSpecName: "kube-api-access-6mbw7") pod "59da35a4-4f0e-4a94-b90d-ff95b13ed602" (UID: "59da35a4-4f0e-4a94-b90d-ff95b13ed602"). InnerVolumeSpecName "kube-api-access-6mbw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.800518 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59da35a4-4f0e-4a94-b90d-ff95b13ed602" (UID: "59da35a4-4f0e-4a94-b90d-ff95b13ed602"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.870646 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mbw7\" (UniqueName: \"kubernetes.io/projected/59da35a4-4f0e-4a94-b90d-ff95b13ed602-kube-api-access-6mbw7\") on node \"crc\" DevicePath \"\"" Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.870677 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:47:34 crc kubenswrapper[4894]: I1208 15:47:34.870686 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59da35a4-4f0e-4a94-b90d-ff95b13ed602-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.108526 4894 generic.go:334] "Generic (PLEG): container finished" podID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerID="7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11" exitCode=0 Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.108594 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerDied","Data":"7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11"} Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.108614 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h7mv4" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.109791 4894 scope.go:117] "RemoveContainer" containerID="7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.109723 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7mv4" event={"ID":"59da35a4-4f0e-4a94-b90d-ff95b13ed602","Type":"ContainerDied","Data":"77794e09b7fa7bbcae7711a66ae4fff6a0c77bdf00d9b6cd24d7e9391a328ea2"} Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.146426 4894 scope.go:117] "RemoveContainer" containerID="483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.152100 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h7mv4"] Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.161251 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h7mv4"] Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.181449 4894 scope.go:117] "RemoveContainer" containerID="1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.212330 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" path="/var/lib/kubelet/pods/59da35a4-4f0e-4a94-b90d-ff95b13ed602/volumes" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.244190 4894 scope.go:117] "RemoveContainer" containerID="7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11" Dec 08 15:47:35 crc kubenswrapper[4894]: E1208 15:47:35.244737 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11\": container with ID starting with 7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11 not found: ID does not exist" containerID="7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.244775 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11"} err="failed to get container status \"7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11\": rpc error: code = NotFound desc = could not find container \"7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11\": container with ID starting with 7fd89e4960a481e32f1c00e1892ac893d5003a82b5960cb3aff2d1c688d6dd11 not found: ID does not exist" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.244801 4894 scope.go:117] "RemoveContainer" containerID="483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36" Dec 08 15:47:35 crc kubenswrapper[4894]: E1208 15:47:35.245417 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36\": container with ID starting with 483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36 not found: ID does not exist" containerID="483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.245450 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36"} err="failed to get container status \"483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36\": rpc error: code = NotFound desc = could not find container \"483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36\": container with ID starting with 483f3cb7df84818b71bbfe294a1e30c111232702061027782e8b6d5e5599dc36 not found: ID does not exist" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.245468 4894 scope.go:117] "RemoveContainer" containerID="1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629" Dec 08 15:47:35 crc kubenswrapper[4894]: E1208 15:47:35.252402 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629\": container with ID starting with 1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629 not found: ID does not exist" containerID="1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629" Dec 08 15:47:35 crc kubenswrapper[4894]: I1208 15:47:35.252458 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629"} err="failed to get container status \"1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629\": rpc error: code = NotFound desc = could not find container \"1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629\": container with ID starting with 1a4d699854280ddae998a865669de0e9e22e7f2809a5588c2a931cff62431629 not found: ID does not exist" Dec 08 15:47:42 crc kubenswrapper[4894]: I1208 15:47:42.197042 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:47:42 crc kubenswrapper[4894]: E1208 15:47:42.197586 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:47:53 crc kubenswrapper[4894]: I1208 15:47:53.305861 4894 generic.go:334] "Generic (PLEG): container finished" podID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerID="ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa" exitCode=0 Dec 08 15:47:53 crc kubenswrapper[4894]: I1208 15:47:53.305975 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c4b76/must-gather-fjsbr" event={"ID":"23675181-ae3e-4a38-9f92-06ed475dfa33","Type":"ContainerDied","Data":"ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa"} Dec 08 15:47:53 crc kubenswrapper[4894]: I1208 15:47:53.307197 4894 scope.go:117] "RemoveContainer" containerID="ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa" Dec 08 15:47:54 crc kubenswrapper[4894]: I1208 15:47:54.090471 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c4b76_must-gather-fjsbr_23675181-ae3e-4a38-9f92-06ed475dfa33/gather/0.log" Dec 08 15:47:56 crc kubenswrapper[4894]: I1208 15:47:56.197173 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:47:56 crc kubenswrapper[4894]: E1208 15:47:56.197712 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:48:01 crc kubenswrapper[4894]: I1208 15:48:01.676853 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c4b76/must-gather-fjsbr"] Dec 08 15:48:01 crc kubenswrapper[4894]: I1208 15:48:01.677782 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-c4b76/must-gather-fjsbr" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerName="copy" containerID="cri-o://575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662" gracePeriod=2 Dec 08 15:48:01 crc kubenswrapper[4894]: I1208 15:48:01.713329 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c4b76/must-gather-fjsbr"] Dec 08 15:48:01 crc kubenswrapper[4894]: E1208 15:48:01.938515 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23675181_ae3e_4a38_9f92_06ed475dfa33.slice/crio-575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23675181_ae3e_4a38_9f92_06ed475dfa33.slice/crio-conmon-575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662.scope\": RecentStats: unable to find data in memory cache]" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.111918 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c4b76_must-gather-fjsbr_23675181-ae3e-4a38-9f92-06ed475dfa33/copy/0.log" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.112681 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.127684 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfr5w\" (UniqueName: \"kubernetes.io/projected/23675181-ae3e-4a38-9f92-06ed475dfa33-kube-api-access-vfr5w\") pod \"23675181-ae3e-4a38-9f92-06ed475dfa33\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.127744 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23675181-ae3e-4a38-9f92-06ed475dfa33-must-gather-output\") pod \"23675181-ae3e-4a38-9f92-06ed475dfa33\" (UID: \"23675181-ae3e-4a38-9f92-06ed475dfa33\") " Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.136768 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23675181-ae3e-4a38-9f92-06ed475dfa33-kube-api-access-vfr5w" (OuterVolumeSpecName: "kube-api-access-vfr5w") pod "23675181-ae3e-4a38-9f92-06ed475dfa33" (UID: "23675181-ae3e-4a38-9f92-06ed475dfa33"). InnerVolumeSpecName "kube-api-access-vfr5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.230979 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfr5w\" (UniqueName: \"kubernetes.io/projected/23675181-ae3e-4a38-9f92-06ed475dfa33-kube-api-access-vfr5w\") on node \"crc\" DevicePath \"\"" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.261918 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23675181-ae3e-4a38-9f92-06ed475dfa33-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "23675181-ae3e-4a38-9f92-06ed475dfa33" (UID: "23675181-ae3e-4a38-9f92-06ed475dfa33"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.333135 4894 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23675181-ae3e-4a38-9f92-06ed475dfa33-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.402848 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c4b76_must-gather-fjsbr_23675181-ae3e-4a38-9f92-06ed475dfa33/copy/0.log" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.403466 4894 generic.go:334] "Generic (PLEG): container finished" podID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerID="575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662" exitCode=143 Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.403512 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c4b76/must-gather-fjsbr" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.403527 4894 scope.go:117] "RemoveContainer" containerID="575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.430703 4894 scope.go:117] "RemoveContainer" containerID="ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.508714 4894 scope.go:117] "RemoveContainer" containerID="575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662" Dec 08 15:48:02 crc kubenswrapper[4894]: E1208 15:48:02.509301 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662\": container with ID starting with 575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662 not found: ID does not exist" containerID="575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.509339 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662"} err="failed to get container status \"575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662\": rpc error: code = NotFound desc = could not find container \"575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662\": container with ID starting with 575dee1cf795a7a9e07a368905cf2dfc51a427b25966f4626aa33dda0cd34662 not found: ID does not exist" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.509384 4894 scope.go:117] "RemoveContainer" containerID="ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa" Dec 08 15:48:02 crc kubenswrapper[4894]: E1208 15:48:02.509837 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa\": container with ID starting with ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa not found: ID does not exist" containerID="ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa" Dec 08 15:48:02 crc kubenswrapper[4894]: I1208 15:48:02.509863 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa"} err="failed to get container status \"ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa\": rpc error: code = NotFound desc = could not find container \"ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa\": container with ID starting with ec54875029e7d2460d7769fb7d8b3ce3990708695505a195ef2cd965276488fa not found: ID does not exist" Dec 08 15:48:03 crc kubenswrapper[4894]: I1208 15:48:03.206114 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" path="/var/lib/kubelet/pods/23675181-ae3e-4a38-9f92-06ed475dfa33/volumes" Dec 08 15:48:07 crc kubenswrapper[4894]: I1208 15:48:07.197538 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:48:07 crc kubenswrapper[4894]: E1208 15:48:07.197884 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:48:19 crc kubenswrapper[4894]: I1208 15:48:19.208472 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:48:19 crc kubenswrapper[4894]: E1208 15:48:19.209425 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:48:30 crc kubenswrapper[4894]: I1208 15:48:30.197900 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:48:30 crc kubenswrapper[4894]: E1208 15:48:30.198880 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:48:41 crc kubenswrapper[4894]: I1208 15:48:41.198942 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:48:41 crc kubenswrapper[4894]: E1208 15:48:41.200590 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:48:52 crc kubenswrapper[4894]: I1208 15:48:52.196880 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:48:52 crc kubenswrapper[4894]: E1208 15:48:52.197889 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:49:07 crc kubenswrapper[4894]: I1208 15:49:07.198150 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:49:07 crc kubenswrapper[4894]: E1208 15:49:07.199348 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:49:20 crc kubenswrapper[4894]: I1208 15:49:20.196568 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:49:20 crc kubenswrapper[4894]: E1208 15:49:20.197316 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:49:35 crc kubenswrapper[4894]: I1208 15:49:35.197174 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:49:35 crc kubenswrapper[4894]: E1208 15:49:35.197990 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:49:47 crc kubenswrapper[4894]: I1208 15:49:47.196517 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:49:47 crc kubenswrapper[4894]: E1208 15:49:47.197338 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:49:59 crc kubenswrapper[4894]: I1208 15:49:59.353326 4894 scope.go:117] "RemoveContainer" containerID="7ff7270ae40210262218f106b0abc030182b20dc5e46072ecf0607300cd59c48" Dec 08 15:50:00 crc kubenswrapper[4894]: I1208 15:50:00.196657 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:50:00 crc kubenswrapper[4894]: E1208 15:50:00.197012 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:50:14 crc kubenswrapper[4894]: I1208 15:50:14.196534 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:50:14 crc kubenswrapper[4894]: E1208 15:50:14.197342 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:50:28 crc kubenswrapper[4894]: I1208 15:50:28.197162 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:50:28 crc kubenswrapper[4894]: E1208 15:50:28.197911 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:50:39 crc kubenswrapper[4894]: I1208 15:50:39.210239 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:50:39 crc kubenswrapper[4894]: E1208 15:50:39.210934 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.442179 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mghdw/must-gather-slfh5"] Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443028 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="extract-utilities" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443040 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="extract-utilities" Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443064 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerName="gather" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443070 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerName="gather" Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443082 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="extract-content" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443089 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="extract-content" Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443105 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerName="copy" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443110 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerName="copy" Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443120 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="registry-server" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443125 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="registry-server" Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443135 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="registry-server" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443140 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="registry-server" Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443153 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="extract-utilities" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443159 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="extract-utilities" Dec 08 15:50:48 crc kubenswrapper[4894]: E1208 15:50:48.443172 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="extract-content" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443178 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="extract-content" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443375 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerName="copy" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443392 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="458c93d2-67fd-4afe-9caf-6364d702d0ad" containerName="registry-server" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443410 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="59da35a4-4f0e-4a94-b90d-ff95b13ed602" containerName="registry-server" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.443420 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="23675181-ae3e-4a38-9f92-06ed475dfa33" containerName="gather" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.444384 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.447138 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mghdw"/"default-dockercfg-fq8qx" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.447226 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mghdw"/"openshift-service-ca.crt" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.447998 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mghdw"/"kube-root-ca.crt" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.464569 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mghdw/must-gather-slfh5"] Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.543316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-must-gather-output\") pod \"must-gather-slfh5\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.543626 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45x9r\" (UniqueName: \"kubernetes.io/projected/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-kube-api-access-45x9r\") pod \"must-gather-slfh5\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.645880 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-must-gather-output\") pod \"must-gather-slfh5\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.645928 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45x9r\" (UniqueName: \"kubernetes.io/projected/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-kube-api-access-45x9r\") pod \"must-gather-slfh5\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.646351 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-must-gather-output\") pod \"must-gather-slfh5\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.678518 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45x9r\" (UniqueName: \"kubernetes.io/projected/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-kube-api-access-45x9r\") pod \"must-gather-slfh5\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:48 crc kubenswrapper[4894]: I1208 15:50:48.768222 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:50:49 crc kubenswrapper[4894]: I1208 15:50:49.286356 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mghdw/must-gather-slfh5"] Dec 08 15:50:49 crc kubenswrapper[4894]: I1208 15:50:49.790677 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/must-gather-slfh5" event={"ID":"6352c3c4-cc30-4dbc-ba45-c20e92b7a782","Type":"ContainerStarted","Data":"37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10"} Dec 08 15:50:49 crc kubenswrapper[4894]: I1208 15:50:49.790720 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/must-gather-slfh5" event={"ID":"6352c3c4-cc30-4dbc-ba45-c20e92b7a782","Type":"ContainerStarted","Data":"b91b8e69d3d13c1726e843f0c70765dc4eeb250d1e13976efafc30942732695f"} Dec 08 15:50:50 crc kubenswrapper[4894]: I1208 15:50:50.803162 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/must-gather-slfh5" event={"ID":"6352c3c4-cc30-4dbc-ba45-c20e92b7a782","Type":"ContainerStarted","Data":"ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7"} Dec 08 15:50:50 crc kubenswrapper[4894]: I1208 15:50:50.827168 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mghdw/must-gather-slfh5" podStartSLOduration=2.827139596 podStartE2EDuration="2.827139596s" podCreationTimestamp="2025-12-08 15:50:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:50:50.820723935 +0000 UTC m=+3871.920730050" watchObservedRunningTime="2025-12-08 15:50:50.827139596 +0000 UTC m=+3871.927145711" Dec 08 15:50:51 crc kubenswrapper[4894]: I1208 15:50:51.840410 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gtv76"] Dec 08 15:50:51 crc kubenswrapper[4894]: I1208 15:50:51.843754 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:51 crc kubenswrapper[4894]: I1208 15:50:51.856125 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gtv76"] Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.019143 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-catalog-content\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.019564 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-utilities\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.020073 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwnp7\" (UniqueName: \"kubernetes.io/projected/2b469c3c-6d2b-467c-88d6-055eaff5415d-kube-api-access-vwnp7\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.121345 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwnp7\" (UniqueName: \"kubernetes.io/projected/2b469c3c-6d2b-467c-88d6-055eaff5415d-kube-api-access-vwnp7\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.121426 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-catalog-content\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.121460 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-utilities\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.121997 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-catalog-content\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.122018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-utilities\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.140193 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwnp7\" (UniqueName: \"kubernetes.io/projected/2b469c3c-6d2b-467c-88d6-055eaff5415d-kube-api-access-vwnp7\") pod \"certified-operators-gtv76\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.181320 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.197842 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:50:52 crc kubenswrapper[4894]: E1208 15:50:52.198325 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.727492 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gtv76"] Dec 08 15:50:52 crc kubenswrapper[4894]: W1208 15:50:52.732563 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b469c3c_6d2b_467c_88d6_055eaff5415d.slice/crio-4bf516f566d7c80d1d5865210c77c8b3cf1c036746f6b45edf8bed304650795d WatchSource:0}: Error finding container 4bf516f566d7c80d1d5865210c77c8b3cf1c036746f6b45edf8bed304650795d: Status 404 returned error can't find the container with id 4bf516f566d7c80d1d5865210c77c8b3cf1c036746f6b45edf8bed304650795d Dec 08 15:50:52 crc kubenswrapper[4894]: I1208 15:50:52.824666 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtv76" event={"ID":"2b469c3c-6d2b-467c-88d6-055eaff5415d","Type":"ContainerStarted","Data":"4bf516f566d7c80d1d5865210c77c8b3cf1c036746f6b45edf8bed304650795d"} Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.637623 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mghdw/crc-debug-72lpj"] Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.639106 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.762323 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-host\") pod \"crc-debug-72lpj\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.762402 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njcws\" (UniqueName: \"kubernetes.io/projected/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-kube-api-access-njcws\") pod \"crc-debug-72lpj\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.833655 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerID="af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee" exitCode=0 Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.833701 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtv76" event={"ID":"2b469c3c-6d2b-467c-88d6-055eaff5415d","Type":"ContainerDied","Data":"af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee"} Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.835333 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.864182 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-host\") pod \"crc-debug-72lpj\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.864256 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njcws\" (UniqueName: \"kubernetes.io/projected/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-kube-api-access-njcws\") pod \"crc-debug-72lpj\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.864660 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-host\") pod \"crc-debug-72lpj\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.886603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njcws\" (UniqueName: \"kubernetes.io/projected/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-kube-api-access-njcws\") pod \"crc-debug-72lpj\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: I1208 15:50:53.956395 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:50:53 crc kubenswrapper[4894]: W1208 15:50:53.985227 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3aacd50_6cb4_4e8e_bbbe_11f9b3ec07ac.slice/crio-c1627eb1f793281ac54cc21c7d7633d73df82b34e814e48de821a82d3d0e5748 WatchSource:0}: Error finding container c1627eb1f793281ac54cc21c7d7633d73df82b34e814e48de821a82d3d0e5748: Status 404 returned error can't find the container with id c1627eb1f793281ac54cc21c7d7633d73df82b34e814e48de821a82d3d0e5748 Dec 08 15:50:54 crc kubenswrapper[4894]: I1208 15:50:54.845942 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtv76" event={"ID":"2b469c3c-6d2b-467c-88d6-055eaff5415d","Type":"ContainerStarted","Data":"21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1"} Dec 08 15:50:54 crc kubenswrapper[4894]: I1208 15:50:54.848003 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/crc-debug-72lpj" event={"ID":"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac","Type":"ContainerStarted","Data":"8c159ce8ca2454be204529d1d5055356358b1b88e04807408c746aa5d6e00d1f"} Dec 08 15:50:54 crc kubenswrapper[4894]: I1208 15:50:54.848045 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/crc-debug-72lpj" event={"ID":"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac","Type":"ContainerStarted","Data":"c1627eb1f793281ac54cc21c7d7633d73df82b34e814e48de821a82d3d0e5748"} Dec 08 15:50:54 crc kubenswrapper[4894]: I1208 15:50:54.885313 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mghdw/crc-debug-72lpj" podStartSLOduration=1.885290564 podStartE2EDuration="1.885290564s" podCreationTimestamp="2025-12-08 15:50:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-08 15:50:54.875780527 +0000 UTC m=+3875.975786642" watchObservedRunningTime="2025-12-08 15:50:54.885290564 +0000 UTC m=+3875.985296679" Dec 08 15:50:55 crc kubenswrapper[4894]: I1208 15:50:55.860193 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerID="21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1" exitCode=0 Dec 08 15:50:55 crc kubenswrapper[4894]: I1208 15:50:55.860426 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtv76" event={"ID":"2b469c3c-6d2b-467c-88d6-055eaff5415d","Type":"ContainerDied","Data":"21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1"} Dec 08 15:50:56 crc kubenswrapper[4894]: I1208 15:50:56.875841 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtv76" event={"ID":"2b469c3c-6d2b-467c-88d6-055eaff5415d","Type":"ContainerStarted","Data":"3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca"} Dec 08 15:50:56 crc kubenswrapper[4894]: I1208 15:50:56.904102 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gtv76" podStartSLOduration=3.354437582 podStartE2EDuration="5.904081277s" podCreationTimestamp="2025-12-08 15:50:51 +0000 UTC" firstStartedPulling="2025-12-08 15:50:53.83514744 +0000 UTC m=+3874.935153545" lastFinishedPulling="2025-12-08 15:50:56.384791125 +0000 UTC m=+3877.484797240" observedRunningTime="2025-12-08 15:50:56.901137465 +0000 UTC m=+3878.001143590" watchObservedRunningTime="2025-12-08 15:50:56.904081277 +0000 UTC m=+3878.004087392" Dec 08 15:50:59 crc kubenswrapper[4894]: I1208 15:50:59.406177 4894 scope.go:117] "RemoveContainer" containerID="a3c64c0e829a54a4a585eeb857858b820ededc4925c8d04836d8f1c798796521" Dec 08 15:51:02 crc kubenswrapper[4894]: I1208 15:51:02.182417 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:51:02 crc kubenswrapper[4894]: I1208 15:51:02.182938 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:51:02 crc kubenswrapper[4894]: I1208 15:51:02.230661 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:51:02 crc kubenswrapper[4894]: I1208 15:51:02.999753 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:51:03 crc kubenswrapper[4894]: I1208 15:51:03.228605 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gtv76"] Dec 08 15:51:04 crc kubenswrapper[4894]: I1208 15:51:04.945366 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gtv76" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="registry-server" containerID="cri-o://3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca" gracePeriod=2 Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.197068 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:51:05 crc kubenswrapper[4894]: E1208 15:51:05.197637 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.808251 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.815328 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-utilities\") pod \"2b469c3c-6d2b-467c-88d6-055eaff5415d\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.815371 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwnp7\" (UniqueName: \"kubernetes.io/projected/2b469c3c-6d2b-467c-88d6-055eaff5415d-kube-api-access-vwnp7\") pod \"2b469c3c-6d2b-467c-88d6-055eaff5415d\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.815588 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-catalog-content\") pod \"2b469c3c-6d2b-467c-88d6-055eaff5415d\" (UID: \"2b469c3c-6d2b-467c-88d6-055eaff5415d\") " Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.817396 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-utilities" (OuterVolumeSpecName: "utilities") pod "2b469c3c-6d2b-467c-88d6-055eaff5415d" (UID: "2b469c3c-6d2b-467c-88d6-055eaff5415d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.822973 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b469c3c-6d2b-467c-88d6-055eaff5415d-kube-api-access-vwnp7" (OuterVolumeSpecName: "kube-api-access-vwnp7") pod "2b469c3c-6d2b-467c-88d6-055eaff5415d" (UID: "2b469c3c-6d2b-467c-88d6-055eaff5415d"). InnerVolumeSpecName "kube-api-access-vwnp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.885070 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b469c3c-6d2b-467c-88d6-055eaff5415d" (UID: "2b469c3c-6d2b-467c-88d6-055eaff5415d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.918281 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.918320 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b469c3c-6d2b-467c-88d6-055eaff5415d-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.918330 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwnp7\" (UniqueName: \"kubernetes.io/projected/2b469c3c-6d2b-467c-88d6-055eaff5415d-kube-api-access-vwnp7\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.955614 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerID="3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca" exitCode=0 Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.955664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtv76" event={"ID":"2b469c3c-6d2b-467c-88d6-055eaff5415d","Type":"ContainerDied","Data":"3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca"} Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.955697 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtv76" event={"ID":"2b469c3c-6d2b-467c-88d6-055eaff5415d","Type":"ContainerDied","Data":"4bf516f566d7c80d1d5865210c77c8b3cf1c036746f6b45edf8bed304650795d"} Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.955719 4894 scope.go:117] "RemoveContainer" containerID="3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.955932 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtv76" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.981625 4894 scope.go:117] "RemoveContainer" containerID="21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1" Dec 08 15:51:05 crc kubenswrapper[4894]: I1208 15:51:05.996160 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gtv76"] Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.011169 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gtv76"] Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.027173 4894 scope.go:117] "RemoveContainer" containerID="af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee" Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.051382 4894 scope.go:117] "RemoveContainer" containerID="3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca" Dec 08 15:51:06 crc kubenswrapper[4894]: E1208 15:51:06.052063 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca\": container with ID starting with 3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca not found: ID does not exist" containerID="3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca" Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.052198 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca"} err="failed to get container status \"3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca\": rpc error: code = NotFound desc = could not find container \"3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca\": container with ID starting with 3d5daec61d039de0127743c09a84f6b60b5b24bc9c9f5f13b8cdd0770ebbcfca not found: ID does not exist" Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.052287 4894 scope.go:117] "RemoveContainer" containerID="21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1" Dec 08 15:51:06 crc kubenswrapper[4894]: E1208 15:51:06.053009 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1\": container with ID starting with 21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1 not found: ID does not exist" containerID="21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1" Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.053063 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1"} err="failed to get container status \"21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1\": rpc error: code = NotFound desc = could not find container \"21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1\": container with ID starting with 21e2fa735c0521984e796980aeaeeac2e18209381d22bfc036873c11e44a7cb1 not found: ID does not exist" Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.053100 4894 scope.go:117] "RemoveContainer" containerID="af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee" Dec 08 15:51:06 crc kubenswrapper[4894]: E1208 15:51:06.055568 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee\": container with ID starting with af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee not found: ID does not exist" containerID="af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee" Dec 08 15:51:06 crc kubenswrapper[4894]: I1208 15:51:06.055620 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee"} err="failed to get container status \"af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee\": rpc error: code = NotFound desc = could not find container \"af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee\": container with ID starting with af36f9cd025513f79a8212051d89454b09d47ebe3c6b9a71fab7f190dcfca8ee not found: ID does not exist" Dec 08 15:51:07 crc kubenswrapper[4894]: I1208 15:51:07.207029 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" path="/var/lib/kubelet/pods/2b469c3c-6d2b-467c-88d6-055eaff5415d/volumes" Dec 08 15:51:16 crc kubenswrapper[4894]: I1208 15:51:16.197858 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:51:16 crc kubenswrapper[4894]: E1208 15:51:16.198539 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:51:30 crc kubenswrapper[4894]: I1208 15:51:30.197773 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:51:30 crc kubenswrapper[4894]: E1208 15:51:30.198595 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:51:31 crc kubenswrapper[4894]: I1208 15:51:31.177521 4894 generic.go:334] "Generic (PLEG): container finished" podID="b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac" containerID="8c159ce8ca2454be204529d1d5055356358b1b88e04807408c746aa5d6e00d1f" exitCode=0 Dec 08 15:51:31 crc kubenswrapper[4894]: I1208 15:51:31.177574 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/crc-debug-72lpj" event={"ID":"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac","Type":"ContainerDied","Data":"8c159ce8ca2454be204529d1d5055356358b1b88e04807408c746aa5d6e00d1f"} Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.743974 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.778668 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mghdw/crc-debug-72lpj"] Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.786662 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mghdw/crc-debug-72lpj"] Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.851668 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njcws\" (UniqueName: \"kubernetes.io/projected/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-kube-api-access-njcws\") pod \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.851709 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-host\") pod \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\" (UID: \"b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac\") " Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.851901 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-host" (OuterVolumeSpecName: "host") pod "b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac" (UID: "b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.852231 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-host\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.866058 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-kube-api-access-njcws" (OuterVolumeSpecName: "kube-api-access-njcws") pod "b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac" (UID: "b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac"). InnerVolumeSpecName "kube-api-access-njcws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:51:32 crc kubenswrapper[4894]: I1208 15:51:32.954508 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njcws\" (UniqueName: \"kubernetes.io/projected/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac-kube-api-access-njcws\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.196530 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-72lpj" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.211494 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac" path="/var/lib/kubelet/pods/b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac/volumes" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.212219 4894 scope.go:117] "RemoveContainer" containerID="8c159ce8ca2454be204529d1d5055356358b1b88e04807408c746aa5d6e00d1f" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.968274 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mghdw/crc-debug-qt284"] Dec 08 15:51:33 crc kubenswrapper[4894]: E1208 15:51:33.968925 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="extract-content" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.968937 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="extract-content" Dec 08 15:51:33 crc kubenswrapper[4894]: E1208 15:51:33.968954 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac" containerName="container-00" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.968961 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac" containerName="container-00" Dec 08 15:51:33 crc kubenswrapper[4894]: E1208 15:51:33.968974 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="registry-server" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.968980 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="registry-server" Dec 08 15:51:33 crc kubenswrapper[4894]: E1208 15:51:33.968993 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="extract-utilities" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.969000 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="extract-utilities" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.969187 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b469c3c-6d2b-467c-88d6-055eaff5415d" containerName="registry-server" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.969198 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3aacd50-6cb4-4e8e-bbbe-11f9b3ec07ac" containerName="container-00" Dec 08 15:51:33 crc kubenswrapper[4894]: I1208 15:51:33.969836 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:34 crc kubenswrapper[4894]: I1208 15:51:34.076592 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtsrq\" (UniqueName: \"kubernetes.io/projected/9727a7c5-0646-4161-be3f-d9f7487bdade-kube-api-access-rtsrq\") pod \"crc-debug-qt284\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:34 crc kubenswrapper[4894]: I1208 15:51:34.077042 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9727a7c5-0646-4161-be3f-d9f7487bdade-host\") pod \"crc-debug-qt284\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:34 crc kubenswrapper[4894]: I1208 15:51:34.183913 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9727a7c5-0646-4161-be3f-d9f7487bdade-host\") pod \"crc-debug-qt284\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:34 crc kubenswrapper[4894]: I1208 15:51:34.184098 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtsrq\" (UniqueName: \"kubernetes.io/projected/9727a7c5-0646-4161-be3f-d9f7487bdade-kube-api-access-rtsrq\") pod \"crc-debug-qt284\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:34 crc kubenswrapper[4894]: I1208 15:51:34.184695 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9727a7c5-0646-4161-be3f-d9f7487bdade-host\") pod \"crc-debug-qt284\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:34 crc kubenswrapper[4894]: I1208 15:51:34.208123 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtsrq\" (UniqueName: \"kubernetes.io/projected/9727a7c5-0646-4161-be3f-d9f7487bdade-kube-api-access-rtsrq\") pod \"crc-debug-qt284\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:34 crc kubenswrapper[4894]: I1208 15:51:34.288065 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:35 crc kubenswrapper[4894]: I1208 15:51:35.220696 4894 generic.go:334] "Generic (PLEG): container finished" podID="9727a7c5-0646-4161-be3f-d9f7487bdade" containerID="45a43105de1d02a7ae13df31273b921c05a5dac0168eb226b58e2d60405cd823" exitCode=0 Dec 08 15:51:35 crc kubenswrapper[4894]: I1208 15:51:35.221213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/crc-debug-qt284" event={"ID":"9727a7c5-0646-4161-be3f-d9f7487bdade","Type":"ContainerDied","Data":"45a43105de1d02a7ae13df31273b921c05a5dac0168eb226b58e2d60405cd823"} Dec 08 15:51:35 crc kubenswrapper[4894]: I1208 15:51:35.221242 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/crc-debug-qt284" event={"ID":"9727a7c5-0646-4161-be3f-d9f7487bdade","Type":"ContainerStarted","Data":"28c2717599e01b094a87a6fd2bed64f4f73e031de392debf54b3f2274fcfe184"} Dec 08 15:51:35 crc kubenswrapper[4894]: I1208 15:51:35.662420 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mghdw/crc-debug-qt284"] Dec 08 15:51:35 crc kubenswrapper[4894]: I1208 15:51:35.677954 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mghdw/crc-debug-qt284"] Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.345969 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.527170 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9727a7c5-0646-4161-be3f-d9f7487bdade-host\") pod \"9727a7c5-0646-4161-be3f-d9f7487bdade\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.527280 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtsrq\" (UniqueName: \"kubernetes.io/projected/9727a7c5-0646-4161-be3f-d9f7487bdade-kube-api-access-rtsrq\") pod \"9727a7c5-0646-4161-be3f-d9f7487bdade\" (UID: \"9727a7c5-0646-4161-be3f-d9f7487bdade\") " Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.528517 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9727a7c5-0646-4161-be3f-d9f7487bdade-host" (OuterVolumeSpecName: "host") pod "9727a7c5-0646-4161-be3f-d9f7487bdade" (UID: "9727a7c5-0646-4161-be3f-d9f7487bdade"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.533855 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9727a7c5-0646-4161-be3f-d9f7487bdade-kube-api-access-rtsrq" (OuterVolumeSpecName: "kube-api-access-rtsrq") pod "9727a7c5-0646-4161-be3f-d9f7487bdade" (UID: "9727a7c5-0646-4161-be3f-d9f7487bdade"). InnerVolumeSpecName "kube-api-access-rtsrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.630007 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9727a7c5-0646-4161-be3f-d9f7487bdade-host\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.630584 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtsrq\" (UniqueName: \"kubernetes.io/projected/9727a7c5-0646-4161-be3f-d9f7487bdade-kube-api-access-rtsrq\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.863628 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mghdw/crc-debug-wmnzs"] Dec 08 15:51:36 crc kubenswrapper[4894]: E1208 15:51:36.864022 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9727a7c5-0646-4161-be3f-d9f7487bdade" containerName="container-00" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.864040 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9727a7c5-0646-4161-be3f-d9f7487bdade" containerName="container-00" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.864229 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9727a7c5-0646-4161-be3f-d9f7487bdade" containerName="container-00" Dec 08 15:51:36 crc kubenswrapper[4894]: I1208 15:51:36.864778 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.037961 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a924c43-b351-4568-a1e5-963430ea1e31-host\") pod \"crc-debug-wmnzs\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.038025 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svtlq\" (UniqueName: \"kubernetes.io/projected/2a924c43-b351-4568-a1e5-963430ea1e31-kube-api-access-svtlq\") pod \"crc-debug-wmnzs\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.140297 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a924c43-b351-4568-a1e5-963430ea1e31-host\") pod \"crc-debug-wmnzs\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.140338 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svtlq\" (UniqueName: \"kubernetes.io/projected/2a924c43-b351-4568-a1e5-963430ea1e31-kube-api-access-svtlq\") pod \"crc-debug-wmnzs\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.140435 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a924c43-b351-4568-a1e5-963430ea1e31-host\") pod \"crc-debug-wmnzs\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.169718 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svtlq\" (UniqueName: \"kubernetes.io/projected/2a924c43-b351-4568-a1e5-963430ea1e31-kube-api-access-svtlq\") pod \"crc-debug-wmnzs\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.186276 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.214853 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9727a7c5-0646-4161-be3f-d9f7487bdade" path="/var/lib/kubelet/pods/9727a7c5-0646-4161-be3f-d9f7487bdade/volumes" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.245981 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-qt284" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.246028 4894 scope.go:117] "RemoveContainer" containerID="45a43105de1d02a7ae13df31273b921c05a5dac0168eb226b58e2d60405cd823" Dec 08 15:51:37 crc kubenswrapper[4894]: I1208 15:51:37.249015 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/crc-debug-wmnzs" event={"ID":"2a924c43-b351-4568-a1e5-963430ea1e31","Type":"ContainerStarted","Data":"8ad5a3442f3036c24c4e32924278549a27acc930c5980b3ab169d6d6c6e11458"} Dec 08 15:51:37 crc kubenswrapper[4894]: E1208 15:51:37.389376 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9727a7c5_0646_4161_be3f_d9f7487bdade.slice\": RecentStats: unable to find data in memory cache]" Dec 08 15:51:38 crc kubenswrapper[4894]: I1208 15:51:38.261234 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a924c43-b351-4568-a1e5-963430ea1e31" containerID="c9ecf1e32083941c155ca8e06fbc9fa015618f13a20fb083fbf527f4fd3d2d34" exitCode=0 Dec 08 15:51:38 crc kubenswrapper[4894]: I1208 15:51:38.261341 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/crc-debug-wmnzs" event={"ID":"2a924c43-b351-4568-a1e5-963430ea1e31","Type":"ContainerDied","Data":"c9ecf1e32083941c155ca8e06fbc9fa015618f13a20fb083fbf527f4fd3d2d34"} Dec 08 15:51:38 crc kubenswrapper[4894]: I1208 15:51:38.296936 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mghdw/crc-debug-wmnzs"] Dec 08 15:51:38 crc kubenswrapper[4894]: I1208 15:51:38.309931 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mghdw/crc-debug-wmnzs"] Dec 08 15:51:39 crc kubenswrapper[4894]: I1208 15:51:39.375426 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:39 crc kubenswrapper[4894]: I1208 15:51:39.477890 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a924c43-b351-4568-a1e5-963430ea1e31-host\") pod \"2a924c43-b351-4568-a1e5-963430ea1e31\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " Dec 08 15:51:39 crc kubenswrapper[4894]: I1208 15:51:39.477994 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a924c43-b351-4568-a1e5-963430ea1e31-host" (OuterVolumeSpecName: "host") pod "2a924c43-b351-4568-a1e5-963430ea1e31" (UID: "2a924c43-b351-4568-a1e5-963430ea1e31"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 08 15:51:39 crc kubenswrapper[4894]: I1208 15:51:39.478091 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svtlq\" (UniqueName: \"kubernetes.io/projected/2a924c43-b351-4568-a1e5-963430ea1e31-kube-api-access-svtlq\") pod \"2a924c43-b351-4568-a1e5-963430ea1e31\" (UID: \"2a924c43-b351-4568-a1e5-963430ea1e31\") " Dec 08 15:51:39 crc kubenswrapper[4894]: I1208 15:51:39.478482 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a924c43-b351-4568-a1e5-963430ea1e31-host\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:39 crc kubenswrapper[4894]: I1208 15:51:39.484607 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a924c43-b351-4568-a1e5-963430ea1e31-kube-api-access-svtlq" (OuterVolumeSpecName: "kube-api-access-svtlq") pod "2a924c43-b351-4568-a1e5-963430ea1e31" (UID: "2a924c43-b351-4568-a1e5-963430ea1e31"). InnerVolumeSpecName "kube-api-access-svtlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:51:39 crc kubenswrapper[4894]: I1208 15:51:39.580443 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svtlq\" (UniqueName: \"kubernetes.io/projected/2a924c43-b351-4568-a1e5-963430ea1e31-kube-api-access-svtlq\") on node \"crc\" DevicePath \"\"" Dec 08 15:51:40 crc kubenswrapper[4894]: I1208 15:51:40.276558 4894 scope.go:117] "RemoveContainer" containerID="c9ecf1e32083941c155ca8e06fbc9fa015618f13a20fb083fbf527f4fd3d2d34" Dec 08 15:51:40 crc kubenswrapper[4894]: I1208 15:51:40.276749 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/crc-debug-wmnzs" Dec 08 15:51:41 crc kubenswrapper[4894]: I1208 15:51:41.206782 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a924c43-b351-4568-a1e5-963430ea1e31" path="/var/lib/kubelet/pods/2a924c43-b351-4568-a1e5-963430ea1e31/volumes" Dec 08 15:51:42 crc kubenswrapper[4894]: I1208 15:51:42.198616 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:51:43 crc kubenswrapper[4894]: I1208 15:51:43.315122 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"c4eaee5f07322bf8760fc41d72edcb270ed44be93caa67f25d7cb3c1751982eb"} Dec 08 15:52:00 crc kubenswrapper[4894]: I1208 15:52:00.435178 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-789cf67ffd-s9792_7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5/barbican-api/0.log" Dec 08 15:52:00 crc kubenswrapper[4894]: I1208 15:52:00.582680 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-789cf67ffd-s9792_7c1c1c34-52bc-4425-9f7d-75ef02f6e5a5/barbican-api-log/0.log" Dec 08 15:52:00 crc kubenswrapper[4894]: I1208 15:52:00.654285 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5897b758c4-ntx2t_561cc404-97e2-4808-9c11-e83efcb00484/barbican-keystone-listener-log/0.log" Dec 08 15:52:00 crc kubenswrapper[4894]: I1208 15:52:00.665890 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5897b758c4-ntx2t_561cc404-97e2-4808-9c11-e83efcb00484/barbican-keystone-listener/0.log" Dec 08 15:52:00 crc kubenswrapper[4894]: I1208 15:52:00.826734 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-845d6bc787-swrrt_f3b10111-9cb4-4baa-863b-3c34341b2e07/barbican-worker-log/0.log" Dec 08 15:52:00 crc kubenswrapper[4894]: I1208 15:52:00.847297 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-845d6bc787-swrrt_f3b10111-9cb4-4baa-863b-3c34341b2e07/barbican-worker/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.036397 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-8qxph_86e41fca-6958-4b8a-84c0-851b75f8efea/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.051720 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/ceilometer-central-agent/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.114769 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/ceilometer-notification-agent/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.212173 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/proxy-httpd/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.268542 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_4463b2d0-71f4-4554-a73d-7ea950f87081/sg-core/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.367318 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_11765aa0-3b20-4d49-9017-8a56603c10b0/cinder-api/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.418894 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_11765aa0-3b20-4d49-9017-8a56603c10b0/cinder-api-log/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.564270 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f2435a2b-5129-4479-b888-d3874b125874/cinder-scheduler/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.609865 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f2435a2b-5129-4479-b888-d3874b125874/probe/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.759870 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-x999n_e6f24a78-43c1-491e-98d0-3bd4d1d045eb/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.816478 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kqmdx_0f36f23c-cd37-4fe4-b851-6f1aa2c4017a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:01 crc kubenswrapper[4894]: I1208 15:52:01.980108 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-nb62g_487bbd0b-1201-4a53-b475-86f85f5a7f51/init/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.188895 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-7kh7z_d2537804-7d9e-41e2-9224-58916974c998/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.193030 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-nb62g_487bbd0b-1201-4a53-b475-86f85f5a7f51/init/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.211909 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-nb62g_487bbd0b-1201-4a53-b475-86f85f5a7f51/dnsmasq-dns/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.368397 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2d3952ab-f60a-4450-bd40-27b2b99a3e9e/glance-log/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.387130 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2d3952ab-f60a-4450-bd40-27b2b99a3e9e/glance-httpd/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.571370 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_62340785-851a-479a-8c14-3908972a9a9f/glance-httpd/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.643100 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_62340785-851a-479a-8c14-3908972a9a9f/glance-log/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.738336 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dcfbfb8-294k4_c3009b99-5e3b-4f39-b62f-0e04aeb371f2/horizon/0.log" Dec 08 15:52:02 crc kubenswrapper[4894]: I1208 15:52:02.871483 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-h4g7v_d2fb1a31-a297-40a3-9dd3-24e3dc9f58aa/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:03 crc kubenswrapper[4894]: I1208 15:52:03.051461 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-797dcfbfb8-294k4_c3009b99-5e3b-4f39-b62f-0e04aeb371f2/horizon-log/0.log" Dec 08 15:52:03 crc kubenswrapper[4894]: I1208 15:52:03.126726 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-vcccf_dff601df-c402-489d-b878-a8aaa7a32909/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:03 crc kubenswrapper[4894]: I1208 15:52:03.330534 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_5ea146ea-5684-46f7-850c-77df5b66c07b/kube-state-metrics/0.log" Dec 08 15:52:03 crc kubenswrapper[4894]: I1208 15:52:03.399524 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7cb79b8b98-42mc8_57827c7d-7ca8-4eb4-80e9-c6690fca1dac/keystone-api/0.log" Dec 08 15:52:03 crc kubenswrapper[4894]: I1208 15:52:03.564256 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-zr2d7_12612e53-bc51-4b09-9f07-f00d8debac18/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:03 crc kubenswrapper[4894]: I1208 15:52:03.840332 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5576bf896f-fxztt_70163272-391d-4c33-be21-3e504f9a179f/neutron-httpd/0.log" Dec 08 15:52:03 crc kubenswrapper[4894]: I1208 15:52:03.887711 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5576bf896f-fxztt_70163272-391d-4c33-be21-3e504f9a179f/neutron-api/0.log" Dec 08 15:52:04 crc kubenswrapper[4894]: I1208 15:52:04.119898 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7964c_35d2beec-8391-420e-a413-44fed1095880/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:04 crc kubenswrapper[4894]: I1208 15:52:04.607622 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_635376cd-70d9-4108-942a-927f52ce27cb/nova-api-log/0.log" Dec 08 15:52:04 crc kubenswrapper[4894]: I1208 15:52:04.629990 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_96255803-4336-4b64-8aee-956302d0c5a2/nova-cell0-conductor-conductor/0.log" Dec 08 15:52:04 crc kubenswrapper[4894]: I1208 15:52:04.893976 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_635376cd-70d9-4108-942a-927f52ce27cb/nova-api-api/0.log" Dec 08 15:52:05 crc kubenswrapper[4894]: I1208 15:52:05.106191 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e939bafe-db68-41d7-8505-651b08bd1ffd/nova-cell1-conductor-conductor/0.log" Dec 08 15:52:05 crc kubenswrapper[4894]: I1208 15:52:05.193076 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1aee5a31-802a-4d05-b242-a7228bc76fbf/nova-cell1-novncproxy-novncproxy/0.log" Dec 08 15:52:05 crc kubenswrapper[4894]: I1208 15:52:05.344597 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ktkvz_01439e20-f7a9-4aa6-bcd7-754e48a7d8b2/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:05 crc kubenswrapper[4894]: I1208 15:52:05.527336 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097/nova-metadata-log/0.log" Dec 08 15:52:05 crc kubenswrapper[4894]: I1208 15:52:05.807160 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1e121dc9-bd10-476c-b28b-06c6dcb09165/mysql-bootstrap/0.log" Dec 08 15:52:05 crc kubenswrapper[4894]: I1208 15:52:05.810457 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fc76d242-0072-4bef-a299-0a88624cb734/nova-scheduler-scheduler/0.log" Dec 08 15:52:05 crc kubenswrapper[4894]: I1208 15:52:05.975901 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1e121dc9-bd10-476c-b28b-06c6dcb09165/mysql-bootstrap/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.059901 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1e121dc9-bd10-476c-b28b-06c6dcb09165/galera/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.237415 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_755a6729-c028-4206-abc7-5b4bd3825a3e/mysql-bootstrap/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.412253 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_755a6729-c028-4206-abc7-5b4bd3825a3e/galera/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.503646 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_755a6729-c028-4206-abc7-5b4bd3825a3e/mysql-bootstrap/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.642642 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_29e0f64a-aac1-4909-b7f6-d83223c04d31/openstackclient/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.668158 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-89p8p_9d8ef054-d4aa-482b-abad-c961a2a5f2ad/ovn-controller/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.759360 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5dee1e03-ab11-4e3c-b2f2-dd07ca4d4097/nova-metadata-metadata/0.log" Dec 08 15:52:06 crc kubenswrapper[4894]: I1208 15:52:06.873192 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-j22jr_2150dea7-8071-418c-ba2d-20eeb73f2ff8/openstack-network-exporter/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.007256 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovsdb-server-init/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.158159 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovs-vswitchd/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.198141 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovsdb-server-init/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.249005 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zxxl4_e1996355-2ff3-49de-b526-a5cef3dfc253/ovsdb-server/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.346532 4894 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod9727a7c5-0646-4161-be3f-d9f7487bdade"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod9727a7c5-0646-4161-be3f-d9f7487bdade] : Timed out while waiting for systemd to remove kubepods-besteffort-pod9727a7c5_0646_4161_be3f_d9f7487bdade.slice" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.411288 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-fvbzv_0ccc212f-4efb-4444-9156-27751b81562e/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.476901 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_65f9ac8b-3288-4ea6-abc7-c396fb830f69/openstack-network-exporter/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.532065 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_65f9ac8b-3288-4ea6-abc7-c396fb830f69/ovn-northd/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.636986 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_0949aa02-5e23-4bd9-9a93-180a59f701ad/openstack-network-exporter/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.730147 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_0949aa02-5e23-4bd9-9a93-180a59f701ad/ovsdbserver-nb/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.898193 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_31bac65c-93e6-4816-837d-a0f467bb3966/openstack-network-exporter/0.log" Dec 08 15:52:07 crc kubenswrapper[4894]: I1208 15:52:07.947847 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_31bac65c-93e6-4816-837d-a0f467bb3966/ovsdbserver-sb/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.126237 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f74bf4b48-xdrm8_cabb5ceb-58b0-47b5-8db0-cb717c6d8aad/placement-api/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.238480 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f74bf4b48-xdrm8_cabb5ceb-58b0-47b5-8db0-cb717c6d8aad/placement-log/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.254983 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ba215854-1f7f-4315-acf0-edd51b2358cb/setup-container/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.501220 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ba215854-1f7f-4315-acf0-edd51b2358cb/setup-container/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.569628 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8447e7e6-4a92-432c-8a5f-48a4cd10675b/setup-container/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.570395 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ba215854-1f7f-4315-acf0-edd51b2358cb/rabbitmq/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.692642 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8447e7e6-4a92-432c-8a5f-48a4cd10675b/setup-container/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.801117 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8447e7e6-4a92-432c-8a5f-48a4cd10675b/rabbitmq/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.804826 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-glbdv_fae74a7d-767f-47a9-8f3c-1337c1be1d3e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:08 crc kubenswrapper[4894]: I1208 15:52:08.963968 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qsrmq_0147e712-181c-4b8e-85d3-3b90c985f5ee/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:09 crc kubenswrapper[4894]: I1208 15:52:09.062496 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-hnjt7_211799eb-bf07-4965-8404-ec3ddba08c3b/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:09 crc kubenswrapper[4894]: I1208 15:52:09.780472 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lxvp6_f840f001-3515-40c7-bc6e-6434ee8df42a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:09 crc kubenswrapper[4894]: I1208 15:52:09.846200 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-m87rg_b5b86fe7-04aa-4538-88d1-d6dd42673b07/ssh-known-hosts-edpm-deployment/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.073006 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-567d6f4bf7-xrnjd_c7a0f531-0d8b-4858-aee9-981bdfbc81e0/proxy-server/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.100221 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m4f8h_daf51f62-aa82-474b-b3d4-4746043cf399/swift-ring-rebalance/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.204346 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-567d6f4bf7-xrnjd_c7a0f531-0d8b-4858-aee9-981bdfbc81e0/proxy-httpd/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.325451 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-auditor/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.408488 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-reaper/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.484791 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-replicator/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.535616 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-auditor/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.557052 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/account-server/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.684809 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-server/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.699683 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-replicator/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.736149 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/container-updater/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.811061 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-auditor/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.879351 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-expirer/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.918281 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-replicator/0.log" Dec 08 15:52:10 crc kubenswrapper[4894]: I1208 15:52:10.956417 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-server/0.log" Dec 08 15:52:11 crc kubenswrapper[4894]: I1208 15:52:11.053580 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/object-updater/0.log" Dec 08 15:52:11 crc kubenswrapper[4894]: I1208 15:52:11.086837 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/rsync/0.log" Dec 08 15:52:11 crc kubenswrapper[4894]: I1208 15:52:11.094504 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_5e8bdec4-ff8a-4db5-b883-68a8b00328ed/swift-recon-cron/0.log" Dec 08 15:52:11 crc kubenswrapper[4894]: I1208 15:52:11.649774 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-dm9lp_1a77045a-f1f8-46b9-adfa-ad5cfa3c35d6/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:11 crc kubenswrapper[4894]: I1208 15:52:11.650544 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_39de6187-9721-4687-8001-ef13c4abafb3/tempest-tests-tempest-tests-runner/0.log" Dec 08 15:52:11 crc kubenswrapper[4894]: I1208 15:52:11.865962 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_1b11a307-3d6a-4f5e-82eb-b864ef5ed5cb/test-operator-logs-container/0.log" Dec 08 15:52:11 crc kubenswrapper[4894]: I1208 15:52:11.988006 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-s6ppj_18936dba-63ba-44de-954c-c65df5b5743a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 08 15:52:19 crc kubenswrapper[4894]: I1208 15:52:19.974945 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_637635c4-70df-45dd-8139-9c8dc3f53ea0/memcached/0.log" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.285228 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-scwjb"] Dec 08 15:52:30 crc kubenswrapper[4894]: E1208 15:52:30.286317 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a924c43-b351-4568-a1e5-963430ea1e31" containerName="container-00" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.286334 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a924c43-b351-4568-a1e5-963430ea1e31" containerName="container-00" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.286588 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a924c43-b351-4568-a1e5-963430ea1e31" containerName="container-00" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.288586 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.296360 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-scwjb"] Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.402059 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-utilities\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.402103 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjdlj\" (UniqueName: \"kubernetes.io/projected/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-kube-api-access-wjdlj\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.402409 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-catalog-content\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.504586 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-utilities\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.504947 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjdlj\" (UniqueName: \"kubernetes.io/projected/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-kube-api-access-wjdlj\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.505073 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-catalog-content\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.505235 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-utilities\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.505588 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-catalog-content\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.524576 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjdlj\" (UniqueName: \"kubernetes.io/projected/eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736-kube-api-access-wjdlj\") pod \"community-operators-scwjb\" (UID: \"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736\") " pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:30 crc kubenswrapper[4894]: I1208 15:52:30.612793 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:31 crc kubenswrapper[4894]: I1208 15:52:31.176173 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-scwjb"] Dec 08 15:52:31 crc kubenswrapper[4894]: I1208 15:52:31.730973 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scwjb" event={"ID":"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736","Type":"ContainerStarted","Data":"6a9edf4a2727545e641d21e77d920549a76280aed0c564abee25f60ee5859373"} Dec 08 15:52:32 crc kubenswrapper[4894]: I1208 15:52:32.742849 4894 generic.go:334] "Generic (PLEG): container finished" podID="eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736" containerID="5d7afab23c12dca5b3927195b149580892f08571f6fdd4428df588d37ba62a4b" exitCode=0 Dec 08 15:52:32 crc kubenswrapper[4894]: I1208 15:52:32.742913 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scwjb" event={"ID":"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736","Type":"ContainerDied","Data":"5d7afab23c12dca5b3927195b149580892f08571f6fdd4428df588d37ba62a4b"} Dec 08 15:52:36 crc kubenswrapper[4894]: I1208 15:52:36.775587 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scwjb" event={"ID":"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736","Type":"ContainerStarted","Data":"9a69b5f520b06cf7e5e5f25e2510b248a133b9ba9de2ff97e1f322f6b176831f"} Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.221355 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-kh5v9_28cfecb9-9c7c-49ae-af33-fc2d9acd374a/kube-rbac-proxy/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.308978 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-kh5v9_28cfecb9-9c7c-49ae-af33-fc2d9acd374a/manager/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.482412 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8vb7c_4ee4db87-8c97-4002-844b-6c55e1a41b2e/kube-rbac-proxy/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.525270 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-8vb7c_4ee4db87-8c97-4002-844b-6c55e1a41b2e/manager/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.673961 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-f67zc_e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72/kube-rbac-proxy/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.714523 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-f67zc_e6f7ec17-d6e2-4cb2-b1e7-0ad45fae7a72/manager/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.785412 4894 generic.go:334] "Generic (PLEG): container finished" podID="eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736" containerID="9a69b5f520b06cf7e5e5f25e2510b248a133b9ba9de2ff97e1f322f6b176831f" exitCode=0 Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.785461 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scwjb" event={"ID":"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736","Type":"ContainerDied","Data":"9a69b5f520b06cf7e5e5f25e2510b248a133b9ba9de2ff97e1f322f6b176831f"} Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.839656 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/util/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.974261 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/util/0.log" Dec 08 15:52:37 crc kubenswrapper[4894]: I1208 15:52:37.983660 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/pull/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.009939 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/pull/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.158151 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/util/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.183213 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/extract/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.210216 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f5fdb90ec85015bf5fa380e943dc0f460fd3ae178c16d5371da71312744r8wk_11e95246-674f-4e1a-bfce-85a1e76658a1/pull/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.359414 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-ffvpw_b8a908e6-3df2-4dc9-9c9c-068e91dc839c/kube-rbac-proxy/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.383308 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-frz8t_5d7b9bb8-c977-45cb-b198-5f047059a0a4/kube-rbac-proxy/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.466487 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-ffvpw_b8a908e6-3df2-4dc9-9c9c-068e91dc839c/manager/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.562953 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-frz8t_5d7b9bb8-c977-45cb-b198-5f047059a0a4/manager/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.644489 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-mtkrb_afda731d-73a8-443f-9647-a774ca8c3146/kube-rbac-proxy/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.683320 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-mtkrb_afda731d-73a8-443f-9647-a774ca8c3146/manager/0.log" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.795961 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-scwjb" event={"ID":"eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736","Type":"ContainerStarted","Data":"30b05258753f4c59b9ea2267a9ef015c0757012e9c47802042b8c28ce55d65fb"} Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.822643 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-scwjb" podStartSLOduration=3.378687192 podStartE2EDuration="8.822618952s" podCreationTimestamp="2025-12-08 15:52:30 +0000 UTC" firstStartedPulling="2025-12-08 15:52:32.744622391 +0000 UTC m=+3973.844628506" lastFinishedPulling="2025-12-08 15:52:38.188554151 +0000 UTC m=+3979.288560266" observedRunningTime="2025-12-08 15:52:38.813491888 +0000 UTC m=+3979.913498003" watchObservedRunningTime="2025-12-08 15:52:38.822618952 +0000 UTC m=+3979.922625067" Dec 08 15:52:38 crc kubenswrapper[4894]: I1208 15:52:38.830964 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4xrts_c3e89da4-2a46-4982-9392-b0cf87a1edc3/kube-rbac-proxy/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.027503 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-47bpp_f7ec7a43-f5b5-474f-af15-d0e664699477/kube-rbac-proxy/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.048241 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-4xrts_c3e89da4-2a46-4982-9392-b0cf87a1edc3/manager/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.050025 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-47bpp_f7ec7a43-f5b5-474f-af15-d0e664699477/manager/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.263274 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-lbszt_2df93571-3b4e-4ab0-ad8b-34d2ef76034c/kube-rbac-proxy/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.283201 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-lbszt_2df93571-3b4e-4ab0-ad8b-34d2ef76034c/manager/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.437029 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w7xbj_aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2/kube-rbac-proxy/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.483592 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-w7xbj_aa20fa02-6518-4a10-aa2e-2aa4ba28e8f2/manager/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.528696 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-h2xw4_e6585a6c-b50d-41d3-a080-ab7cb27f9199/kube-rbac-proxy/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.833516 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2sfqd_db804bba-5a84-4d23-93ca-e7118afc2af8/kube-rbac-proxy/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.870074 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-h2xw4_e6585a6c-b50d-41d3-a080-ab7cb27f9199/manager/0.log" Dec 08 15:52:39 crc kubenswrapper[4894]: I1208 15:52:39.987297 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2sfqd_db804bba-5a84-4d23-93ca-e7118afc2af8/manager/0.log" Dec 08 15:52:40 crc kubenswrapper[4894]: I1208 15:52:40.107012 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jc2rb_c91237dd-8bb9-4240-b24a-e19b58687d2e/kube-rbac-proxy/0.log" Dec 08 15:52:40 crc kubenswrapper[4894]: I1208 15:52:40.204750 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-jc2rb_c91237dd-8bb9-4240-b24a-e19b58687d2e/manager/0.log" Dec 08 15:52:40 crc kubenswrapper[4894]: I1208 15:52:40.331257 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cqq4p_750c567a-693a-401b-8e7f-d8f1befbeacb/kube-rbac-proxy/0.log" Dec 08 15:52:40 crc kubenswrapper[4894]: I1208 15:52:40.613174 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:40 crc kubenswrapper[4894]: I1208 15:52:40.613493 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:40 crc kubenswrapper[4894]: I1208 15:52:40.823325 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1e121dc9-bd10-476c-b28b-06c6dcb09165" containerName="galera" probeResult="failure" output="command timed out" Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.081261 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4lcw4_fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae/kube-rbac-proxy/0.log" Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.081374 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cqq4p_750c567a-693a-401b-8e7f-d8f1befbeacb/manager/0.log" Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.164788 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879f4lcw4_fa980adc-c90c-4a60-a2b7-e9bfd21ef5ae/manager/0.log" Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.562262 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-647974bbd-56w9g_d6d47a7f-7fcb-4b93-bccd-d6a79fd0e3dd/operator/0.log" Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.632210 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8ptfg_654a1e08-5902-4a0f-8584-f662bc005210/registry-server/0.log" Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.658533 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lr526_2b5f3682-eb12-440c-a016-f165ec4e7cfa/kube-rbac-proxy/0.log" Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.761225 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-scwjb" podUID="eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736" containerName="registry-server" probeResult="failure" output=< Dec 08 15:52:41 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 08 15:52:41 crc kubenswrapper[4894]: > Dec 08 15:52:41 crc kubenswrapper[4894]: I1208 15:52:41.843697 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-lr526_2b5f3682-eb12-440c-a016-f165ec4e7cfa/manager/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.001141 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-sxfkp_59051068-3252-47e6-a826-79d1951c3d5c/manager/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.011662 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-sxfkp_59051068-3252-47e6-a826-79d1951c3d5c/kube-rbac-proxy/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.128974 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-97mkf_9c6f4113-6d2f-4fce-b76b-8802b3b605c6/operator/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.554646 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8695b4b66-c64t7_69d7f3da-48c5-47e8-853b-0cca6246d96e/manager/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.598260 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-69wkl_5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8/kube-rbac-proxy/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.603581 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-6qvtl_2536d7d4-e60f-44a6-8314-cfd34b5545e9/kube-rbac-proxy/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.631297 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-69wkl_5ed01f8d-08b6-4fdf-8478-0c0c373ff3a8/manager/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.797950 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-zhwdn_8b8eda01-4bc4-4eca-b909-901024ef466b/kube-rbac-proxy/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.816208 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-6qvtl_2536d7d4-e60f-44a6-8314-cfd34b5545e9/manager/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.844615 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-zhwdn_8b8eda01-4bc4-4eca-b909-901024ef466b/manager/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.982742 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-rnqzt_8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76/kube-rbac-proxy/0.log" Dec 08 15:52:42 crc kubenswrapper[4894]: I1208 15:52:42.993569 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-rnqzt_8f07e44c-34c8-4ff9-ba05-8ea4b46dfa76/manager/0.log" Dec 08 15:52:50 crc kubenswrapper[4894]: I1208 15:52:50.774337 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:50 crc kubenswrapper[4894]: I1208 15:52:50.836196 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-scwjb" Dec 08 15:52:50 crc kubenswrapper[4894]: I1208 15:52:50.908187 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-scwjb"] Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.013738 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s7cr5"] Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.014045 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s7cr5" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="registry-server" containerID="cri-o://425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807" gracePeriod=2 Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.500089 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.643125 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-catalog-content\") pod \"9393c174-85f4-4267-80a1-e7c1075104b5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.643329 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdpcz\" (UniqueName: \"kubernetes.io/projected/9393c174-85f4-4267-80a1-e7c1075104b5-kube-api-access-mdpcz\") pod \"9393c174-85f4-4267-80a1-e7c1075104b5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.643376 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-utilities\") pod \"9393c174-85f4-4267-80a1-e7c1075104b5\" (UID: \"9393c174-85f4-4267-80a1-e7c1075104b5\") " Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.643919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-utilities" (OuterVolumeSpecName: "utilities") pod "9393c174-85f4-4267-80a1-e7c1075104b5" (UID: "9393c174-85f4-4267-80a1-e7c1075104b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.648760 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9393c174-85f4-4267-80a1-e7c1075104b5-kube-api-access-mdpcz" (OuterVolumeSpecName: "kube-api-access-mdpcz") pod "9393c174-85f4-4267-80a1-e7c1075104b5" (UID: "9393c174-85f4-4267-80a1-e7c1075104b5"). InnerVolumeSpecName "kube-api-access-mdpcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.696208 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9393c174-85f4-4267-80a1-e7c1075104b5" (UID: "9393c174-85f4-4267-80a1-e7c1075104b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.745145 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdpcz\" (UniqueName: \"kubernetes.io/projected/9393c174-85f4-4267-80a1-e7c1075104b5-kube-api-access-mdpcz\") on node \"crc\" DevicePath \"\"" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.745437 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.745531 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9393c174-85f4-4267-80a1-e7c1075104b5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.896645 4894 generic.go:334] "Generic (PLEG): container finished" podID="9393c174-85f4-4267-80a1-e7c1075104b5" containerID="425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807" exitCode=0 Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.896716 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7cr5" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.896772 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7cr5" event={"ID":"9393c174-85f4-4267-80a1-e7c1075104b5","Type":"ContainerDied","Data":"425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807"} Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.898329 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7cr5" event={"ID":"9393c174-85f4-4267-80a1-e7c1075104b5","Type":"ContainerDied","Data":"8fffc084b6575d5e449ccc63e824b1eb0a0f8f9fafdd1ec495226bded68ca452"} Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.898358 4894 scope.go:117] "RemoveContainer" containerID="425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.933839 4894 scope.go:117] "RemoveContainer" containerID="0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae" Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.937573 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s7cr5"] Dec 08 15:52:51 crc kubenswrapper[4894]: I1208 15:52:51.947036 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s7cr5"] Dec 08 15:52:52 crc kubenswrapper[4894]: I1208 15:52:52.537262 4894 scope.go:117] "RemoveContainer" containerID="1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7" Dec 08 15:52:52 crc kubenswrapper[4894]: I1208 15:52:52.594289 4894 scope.go:117] "RemoveContainer" containerID="425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807" Dec 08 15:52:52 crc kubenswrapper[4894]: E1208 15:52:52.594903 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807\": container with ID starting with 425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807 not found: ID does not exist" containerID="425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807" Dec 08 15:52:52 crc kubenswrapper[4894]: I1208 15:52:52.594950 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807"} err="failed to get container status \"425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807\": rpc error: code = NotFound desc = could not find container \"425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807\": container with ID starting with 425cb2deb877ee119140a1bb26c6f44d5fa0c0d6b862c24365362a0396bb7807 not found: ID does not exist" Dec 08 15:52:52 crc kubenswrapper[4894]: I1208 15:52:52.594982 4894 scope.go:117] "RemoveContainer" containerID="0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae" Dec 08 15:52:52 crc kubenswrapper[4894]: E1208 15:52:52.595309 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae\": container with ID starting with 0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae not found: ID does not exist" containerID="0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae" Dec 08 15:52:52 crc kubenswrapper[4894]: I1208 15:52:52.595343 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae"} err="failed to get container status \"0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae\": rpc error: code = NotFound desc = could not find container \"0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae\": container with ID starting with 0db4a3b78b5dd425fcb66c8a803efc8032f2f41ced1210dd2e668f25bc615dae not found: ID does not exist" Dec 08 15:52:52 crc kubenswrapper[4894]: I1208 15:52:52.595366 4894 scope.go:117] "RemoveContainer" containerID="1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7" Dec 08 15:52:52 crc kubenswrapper[4894]: E1208 15:52:52.595767 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7\": container with ID starting with 1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7 not found: ID does not exist" containerID="1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7" Dec 08 15:52:52 crc kubenswrapper[4894]: I1208 15:52:52.595788 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7"} err="failed to get container status \"1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7\": rpc error: code = NotFound desc = could not find container \"1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7\": container with ID starting with 1b804711cfdd77cf2466a836713982a0ca888cdf76f66e2a7db7f375bc07cfd7 not found: ID does not exist" Dec 08 15:52:53 crc kubenswrapper[4894]: I1208 15:52:53.208238 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" path="/var/lib/kubelet/pods/9393c174-85f4-4267-80a1-e7c1075104b5/volumes" Dec 08 15:53:01 crc kubenswrapper[4894]: I1208 15:53:01.100199 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-njrj9_e899bcbb-0ffd-4989-bba4-47b61fe07832/control-plane-machine-set-operator/0.log" Dec 08 15:53:01 crc kubenswrapper[4894]: I1208 15:53:01.575721 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7nbxt_22a82f00-b781-4068-b398-0b4c738db413/kube-rbac-proxy/0.log" Dec 08 15:53:01 crc kubenswrapper[4894]: I1208 15:53:01.589271 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7nbxt_22a82f00-b781-4068-b398-0b4c738db413/machine-api-operator/0.log" Dec 08 15:53:14 crc kubenswrapper[4894]: I1208 15:53:14.134927 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qfw89_960dc174-5cd7-4931-94f7-79871062fcb9/cert-manager-cainjector/0.log" Dec 08 15:53:14 crc kubenswrapper[4894]: I1208 15:53:14.156272 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nzrw5_0aa35d3c-6f53-415d-8a11-f2e8cc6298fb/cert-manager-controller/0.log" Dec 08 15:53:14 crc kubenswrapper[4894]: I1208 15:53:14.301836 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hzvvg_7774fcfd-7cfb-4b62-8e29-48c16c5d0337/cert-manager-webhook/0.log" Dec 08 15:53:25 crc kubenswrapper[4894]: I1208 15:53:25.682788 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5v49c_55c7d1cf-5cb4-4a06-ac50-2ee06f72a522/nmstate-console-plugin/0.log" Dec 08 15:53:25 crc kubenswrapper[4894]: I1208 15:53:25.856391 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-b8g67_2ed9116e-64a3-4bc3-a4fc-50e2a7fd383c/nmstate-handler/0.log" Dec 08 15:53:25 crc kubenswrapper[4894]: I1208 15:53:25.862987 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lkw5q_2f8534ee-633a-4633-9649-dbc49a705529/kube-rbac-proxy/0.log" Dec 08 15:53:25 crc kubenswrapper[4894]: I1208 15:53:25.880097 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-lkw5q_2f8534ee-633a-4633-9649-dbc49a705529/nmstate-metrics/0.log" Dec 08 15:53:26 crc kubenswrapper[4894]: I1208 15:53:26.013143 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-7kjmb_68045b7d-a426-4739-9870-b4e40c54dedd/nmstate-operator/0.log" Dec 08 15:53:26 crc kubenswrapper[4894]: I1208 15:53:26.071998 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-9wlhf_70fa62b1-a4b9-49c3-abd2-49ac96c501b6/nmstate-webhook/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.248120 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6m7rl_e2473db3-776f-4a3d-9b4f-aba30d96463b/kube-rbac-proxy/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.278244 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6m7rl_e2473db3-776f-4a3d-9b4f-aba30d96463b/controller/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.429702 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.617208 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.654573 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.663499 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.692258 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.821228 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.826743 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.831216 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:53:40 crc kubenswrapper[4894]: I1208 15:53:40.908245 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.084149 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-reloader/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.087001 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-frr-files/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.105443 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/cp-metrics/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.119074 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/controller/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.283986 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/frr-metrics/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.338387 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/kube-rbac-proxy-frr/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.382935 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/kube-rbac-proxy/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.490560 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/reloader/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.596868 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-rt5h4_29388ab6-9e49-4662-8d74-1dda6e8d228e/frr-k8s-webhook-server/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.826147 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-78884d4998-whjq5_10bcfd29-cff4-4fcc-be38-43296e67cac7/manager/0.log" Dec 08 15:53:41 crc kubenswrapper[4894]: I1208 15:53:41.996847 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74cdf75695-th2w6_6e5ec2a2-054b-4cb6-a3d6-7376f254869c/webhook-server/0.log" Dec 08 15:53:42 crc kubenswrapper[4894]: I1208 15:53:42.062281 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hxwz6_e316690b-6982-4588-ab38-a33ca86ba1f0/kube-rbac-proxy/0.log" Dec 08 15:53:42 crc kubenswrapper[4894]: I1208 15:53:42.498539 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5r5t2_e7389153-ceef-48e4-a814-b6204ce84627/frr/0.log" Dec 08 15:53:42 crc kubenswrapper[4894]: I1208 15:53:42.581858 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hxwz6_e316690b-6982-4588-ab38-a33ca86ba1f0/speaker/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.098685 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/util/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.284393 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/pull/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.320032 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/pull/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.327599 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/util/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.545619 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/pull/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.574343 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/extract/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.609487 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frdrq6_383ac257-ce50-4d47-8298-799521c308f6/util/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.756537 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/util/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.916129 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/util/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.916778 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/pull/0.log" Dec 08 15:53:55 crc kubenswrapper[4894]: I1208 15:53:55.969289 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/pull/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.097619 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/extract/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.128236 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/pull/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.133839 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83cxg5x_f40ea837-9c2e-4693-960d-246a8ad84e30/util/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.318940 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-utilities/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.495240 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-utilities/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.495517 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-content/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.500284 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-content/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.670694 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-content/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.722622 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/extract-utilities/0.log" Dec 08 15:53:56 crc kubenswrapper[4894]: I1208 15:53:56.905119 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-scwjb_eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736/extract-utilities/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.164232 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-scwjb_eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736/extract-content/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.198262 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-bhxcd_28a7ac74-25ab-4f68-8aee-adbbf5defd99/registry-server/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.222909 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-scwjb_eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736/extract-utilities/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.237079 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-scwjb_eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736/extract-content/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.375549 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-scwjb_eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736/extract-content/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.377513 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-scwjb_eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736/extract-utilities/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.517765 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-scwjb_eaf97a1e-ec51-4ba0-9612-ab0ad5ab4736/registry-server/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.575253 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bl7db_e1fa33df-7fab-40bc-8a30-9004002f770a/marketplace-operator/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.684173 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-utilities/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.906963 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-utilities/0.log" Dec 08 15:53:57 crc kubenswrapper[4894]: I1208 15:53:57.954846 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-content/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.016450 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-content/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.150910 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-utilities/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.194114 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/extract-content/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.337833 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qc7vh_c0e4e418-79da-43c5-8d59-1d1f6a7e2323/registry-server/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.378387 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-utilities/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.503265 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-content/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.511744 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-utilities/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.545294 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-content/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.676029 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-utilities/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.693575 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/extract-content/0.log" Dec 08 15:53:58 crc kubenswrapper[4894]: I1208 15:53:58.996212 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5xhjz_4cbdeb86-106a-426b-a285-ddfe70e7843a/registry-server/0.log" Dec 08 15:54:07 crc kubenswrapper[4894]: I1208 15:54:07.297085 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:54:07 crc kubenswrapper[4894]: I1208 15:54:07.298624 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:54:37 crc kubenswrapper[4894]: I1208 15:54:37.296373 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:54:37 crc kubenswrapper[4894]: I1208 15:54:37.296891 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.296921 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.297535 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.297603 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.299335 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c4eaee5f07322bf8760fc41d72edcb270ed44be93caa67f25d7cb3c1751982eb"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.299522 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://c4eaee5f07322bf8760fc41d72edcb270ed44be93caa67f25d7cb3c1751982eb" gracePeriod=600 Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.909151 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="c4eaee5f07322bf8760fc41d72edcb270ed44be93caa67f25d7cb3c1751982eb" exitCode=0 Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.909555 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"c4eaee5f07322bf8760fc41d72edcb270ed44be93caa67f25d7cb3c1751982eb"} Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.909866 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerStarted","Data":"e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5"} Dec 08 15:55:07 crc kubenswrapper[4894]: I1208 15:55:07.909894 4894 scope.go:117] "RemoveContainer" containerID="dddd6f4802249b53deb6334d7b30e59663c3a786d2f96b08b8e282a68a044295" Dec 08 15:55:40 crc kubenswrapper[4894]: I1208 15:55:40.233996 4894 generic.go:334] "Generic (PLEG): container finished" podID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerID="37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10" exitCode=0 Dec 08 15:55:40 crc kubenswrapper[4894]: I1208 15:55:40.234083 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mghdw/must-gather-slfh5" event={"ID":"6352c3c4-cc30-4dbc-ba45-c20e92b7a782","Type":"ContainerDied","Data":"37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10"} Dec 08 15:55:40 crc kubenswrapper[4894]: I1208 15:55:40.236343 4894 scope.go:117] "RemoveContainer" containerID="37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10" Dec 08 15:55:41 crc kubenswrapper[4894]: I1208 15:55:41.021842 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mghdw_must-gather-slfh5_6352c3c4-cc30-4dbc-ba45-c20e92b7a782/gather/0.log" Dec 08 15:55:51 crc kubenswrapper[4894]: I1208 15:55:51.674798 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mghdw/must-gather-slfh5"] Dec 08 15:55:51 crc kubenswrapper[4894]: I1208 15:55:51.675908 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mghdw/must-gather-slfh5" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerName="copy" containerID="cri-o://ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7" gracePeriod=2 Dec 08 15:55:51 crc kubenswrapper[4894]: I1208 15:55:51.695664 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mghdw/must-gather-slfh5"] Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.121658 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mghdw_must-gather-slfh5_6352c3c4-cc30-4dbc-ba45-c20e92b7a782/copy/0.log" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.122035 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.141339 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45x9r\" (UniqueName: \"kubernetes.io/projected/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-kube-api-access-45x9r\") pod \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.141514 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-must-gather-output\") pod \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\" (UID: \"6352c3c4-cc30-4dbc-ba45-c20e92b7a782\") " Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.152184 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-kube-api-access-45x9r" (OuterVolumeSpecName: "kube-api-access-45x9r") pod "6352c3c4-cc30-4dbc-ba45-c20e92b7a782" (UID: "6352c3c4-cc30-4dbc-ba45-c20e92b7a782"). InnerVolumeSpecName "kube-api-access-45x9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.244007 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45x9r\" (UniqueName: \"kubernetes.io/projected/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-kube-api-access-45x9r\") on node \"crc\" DevicePath \"\"" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.287359 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6352c3c4-cc30-4dbc-ba45-c20e92b7a782" (UID: "6352c3c4-cc30-4dbc-ba45-c20e92b7a782"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.346205 4894 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6352c3c4-cc30-4dbc-ba45-c20e92b7a782-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.356846 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mghdw_must-gather-slfh5_6352c3c4-cc30-4dbc-ba45-c20e92b7a782/copy/0.log" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.357379 4894 generic.go:334] "Generic (PLEG): container finished" podID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerID="ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7" exitCode=143 Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.357430 4894 scope.go:117] "RemoveContainer" containerID="ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.357454 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mghdw/must-gather-slfh5" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.378419 4894 scope.go:117] "RemoveContainer" containerID="37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.651942 4894 scope.go:117] "RemoveContainer" containerID="ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7" Dec 08 15:55:52 crc kubenswrapper[4894]: E1208 15:55:52.652373 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7\": container with ID starting with ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7 not found: ID does not exist" containerID="ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.652411 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7"} err="failed to get container status \"ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7\": rpc error: code = NotFound desc = could not find container \"ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7\": container with ID starting with ff3f7caaeda889b356d882351697b5cad13b7503bcf29a6e3c13d8c835b227e7 not found: ID does not exist" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.652437 4894 scope.go:117] "RemoveContainer" containerID="37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10" Dec 08 15:55:52 crc kubenswrapper[4894]: E1208 15:55:52.652943 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10\": container with ID starting with 37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10 not found: ID does not exist" containerID="37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10" Dec 08 15:55:52 crc kubenswrapper[4894]: I1208 15:55:52.652975 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10"} err="failed to get container status \"37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10\": rpc error: code = NotFound desc = could not find container \"37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10\": container with ID starting with 37f7a69c0f0e99acd9d52f7d7290d7c2bda4130aee5fa2bcc5798cd641917e10 not found: ID does not exist" Dec 08 15:55:53 crc kubenswrapper[4894]: I1208 15:55:53.207613 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" path="/var/lib/kubelet/pods/6352c3c4-cc30-4dbc-ba45-c20e92b7a782/volumes" Dec 08 15:57:07 crc kubenswrapper[4894]: I1208 15:57:07.296997 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:57:07 crc kubenswrapper[4894]: I1208 15:57:07.298478 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:57:37 crc kubenswrapper[4894]: I1208 15:57:37.296367 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:57:37 crc kubenswrapper[4894]: I1208 15:57:37.296891 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.296368 4894 patch_prober.go:28] interesting pod/machine-config-daemon-97dqr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.296889 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.296926 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.297600 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5"} pod="openshift-machine-config-operator/machine-config-daemon-97dqr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.297644 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerName="machine-config-daemon" containerID="cri-o://e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5" gracePeriod=600 Dec 08 15:58:07 crc kubenswrapper[4894]: E1208 15:58:07.423459 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.662097 4894 generic.go:334] "Generic (PLEG): container finished" podID="b27019e5-2a3d-414e-b2ee-7606492ba074" containerID="e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5" exitCode=0 Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.662153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" event={"ID":"b27019e5-2a3d-414e-b2ee-7606492ba074","Type":"ContainerDied","Data":"e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5"} Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.662223 4894 scope.go:117] "RemoveContainer" containerID="c4eaee5f07322bf8760fc41d72edcb270ed44be93caa67f25d7cb3c1751982eb" Dec 08 15:58:07 crc kubenswrapper[4894]: I1208 15:58:07.662901 4894 scope.go:117] "RemoveContainer" containerID="e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5" Dec 08 15:58:07 crc kubenswrapper[4894]: E1208 15:58:07.663237 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:58:18 crc kubenswrapper[4894]: I1208 15:58:18.196830 4894 scope.go:117] "RemoveContainer" containerID="e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5" Dec 08 15:58:18 crc kubenswrapper[4894]: E1208 15:58:18.197642 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.068981 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gnr8s"] Dec 08 15:58:23 crc kubenswrapper[4894]: E1208 15:58:23.069867 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="extract-utilities" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.069882 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="extract-utilities" Dec 08 15:58:23 crc kubenswrapper[4894]: E1208 15:58:23.069903 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerName="copy" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.069909 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerName="copy" Dec 08 15:58:23 crc kubenswrapper[4894]: E1208 15:58:23.069922 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="registry-server" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.069928 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="registry-server" Dec 08 15:58:23 crc kubenswrapper[4894]: E1208 15:58:23.069945 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="extract-content" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.069951 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="extract-content" Dec 08 15:58:23 crc kubenswrapper[4894]: E1208 15:58:23.069963 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerName="gather" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.069969 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerName="gather" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.070157 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9393c174-85f4-4267-80a1-e7c1075104b5" containerName="registry-server" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.070174 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerName="copy" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.070190 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6352c3c4-cc30-4dbc-ba45-c20e92b7a782" containerName="gather" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.071567 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.091830 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnr8s"] Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.273263 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-catalog-content\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.273473 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxvfw\" (UniqueName: \"kubernetes.io/projected/9b39cef3-7078-4992-bd8c-7bbb57193b12-kube-api-access-lxvfw\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.273544 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-utilities\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.375962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxvfw\" (UniqueName: \"kubernetes.io/projected/9b39cef3-7078-4992-bd8c-7bbb57193b12-kube-api-access-lxvfw\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.376057 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-utilities\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.376097 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-catalog-content\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.376629 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-utilities\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.376688 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-catalog-content\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.395843 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxvfw\" (UniqueName: \"kubernetes.io/projected/9b39cef3-7078-4992-bd8c-7bbb57193b12-kube-api-access-lxvfw\") pod \"redhat-marketplace-gnr8s\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:23 crc kubenswrapper[4894]: I1208 15:58:23.437210 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:24 crc kubenswrapper[4894]: I1208 15:58:24.028687 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnr8s"] Dec 08 15:58:24 crc kubenswrapper[4894]: I1208 15:58:24.805825 4894 generic.go:334] "Generic (PLEG): container finished" podID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerID="8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31" exitCode=0 Dec 08 15:58:24 crc kubenswrapper[4894]: I1208 15:58:24.805871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnr8s" event={"ID":"9b39cef3-7078-4992-bd8c-7bbb57193b12","Type":"ContainerDied","Data":"8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31"} Dec 08 15:58:24 crc kubenswrapper[4894]: I1208 15:58:24.805897 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnr8s" event={"ID":"9b39cef3-7078-4992-bd8c-7bbb57193b12","Type":"ContainerStarted","Data":"d51cc11bfb8cfe13b889ea42b0148614c4afe76580d5a2d8107e7de382089bc8"} Dec 08 15:58:24 crc kubenswrapper[4894]: I1208 15:58:24.807826 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 08 15:58:25 crc kubenswrapper[4894]: I1208 15:58:25.816315 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnr8s" event={"ID":"9b39cef3-7078-4992-bd8c-7bbb57193b12","Type":"ContainerStarted","Data":"cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23"} Dec 08 15:58:26 crc kubenswrapper[4894]: I1208 15:58:26.830863 4894 generic.go:334] "Generic (PLEG): container finished" podID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerID="cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23" exitCode=0 Dec 08 15:58:26 crc kubenswrapper[4894]: I1208 15:58:26.830917 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnr8s" event={"ID":"9b39cef3-7078-4992-bd8c-7bbb57193b12","Type":"ContainerDied","Data":"cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23"} Dec 08 15:58:27 crc kubenswrapper[4894]: I1208 15:58:27.840171 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnr8s" event={"ID":"9b39cef3-7078-4992-bd8c-7bbb57193b12","Type":"ContainerStarted","Data":"df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890"} Dec 08 15:58:27 crc kubenswrapper[4894]: I1208 15:58:27.860568 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gnr8s" podStartSLOduration=2.455018716 podStartE2EDuration="4.86055295s" podCreationTimestamp="2025-12-08 15:58:23 +0000 UTC" firstStartedPulling="2025-12-08 15:58:24.807595088 +0000 UTC m=+4325.907601203" lastFinishedPulling="2025-12-08 15:58:27.213129322 +0000 UTC m=+4328.313135437" observedRunningTime="2025-12-08 15:58:27.85704083 +0000 UTC m=+4328.957046955" watchObservedRunningTime="2025-12-08 15:58:27.86055295 +0000 UTC m=+4328.960559065" Dec 08 15:58:32 crc kubenswrapper[4894]: I1208 15:58:32.196629 4894 scope.go:117] "RemoveContainer" containerID="e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5" Dec 08 15:58:32 crc kubenswrapper[4894]: E1208 15:58:32.197382 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:58:33 crc kubenswrapper[4894]: I1208 15:58:33.437563 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:33 crc kubenswrapper[4894]: I1208 15:58:33.438025 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:33 crc kubenswrapper[4894]: I1208 15:58:33.484448 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:33 crc kubenswrapper[4894]: I1208 15:58:33.978785 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:34 crc kubenswrapper[4894]: I1208 15:58:34.036999 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnr8s"] Dec 08 15:58:35 crc kubenswrapper[4894]: I1208 15:58:35.929295 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gnr8s" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="registry-server" containerID="cri-o://df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890" gracePeriod=2 Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.385451 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.549529 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-utilities\") pod \"9b39cef3-7078-4992-bd8c-7bbb57193b12\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.549577 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-catalog-content\") pod \"9b39cef3-7078-4992-bd8c-7bbb57193b12\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.549649 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxvfw\" (UniqueName: \"kubernetes.io/projected/9b39cef3-7078-4992-bd8c-7bbb57193b12-kube-api-access-lxvfw\") pod \"9b39cef3-7078-4992-bd8c-7bbb57193b12\" (UID: \"9b39cef3-7078-4992-bd8c-7bbb57193b12\") " Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.551758 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-utilities" (OuterVolumeSpecName: "utilities") pod "9b39cef3-7078-4992-bd8c-7bbb57193b12" (UID: "9b39cef3-7078-4992-bd8c-7bbb57193b12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.557485 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b39cef3-7078-4992-bd8c-7bbb57193b12-kube-api-access-lxvfw" (OuterVolumeSpecName: "kube-api-access-lxvfw") pod "9b39cef3-7078-4992-bd8c-7bbb57193b12" (UID: "9b39cef3-7078-4992-bd8c-7bbb57193b12"). InnerVolumeSpecName "kube-api-access-lxvfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.584501 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b39cef3-7078-4992-bd8c-7bbb57193b12" (UID: "9b39cef3-7078-4992-bd8c-7bbb57193b12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.652195 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-utilities\") on node \"crc\" DevicePath \"\"" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.652221 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b39cef3-7078-4992-bd8c-7bbb57193b12-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.652233 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxvfw\" (UniqueName: \"kubernetes.io/projected/9b39cef3-7078-4992-bd8c-7bbb57193b12-kube-api-access-lxvfw\") on node \"crc\" DevicePath \"\"" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.938483 4894 generic.go:334] "Generic (PLEG): container finished" podID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerID="df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890" exitCode=0 Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.938533 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnr8s" event={"ID":"9b39cef3-7078-4992-bd8c-7bbb57193b12","Type":"ContainerDied","Data":"df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890"} Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.938569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gnr8s" event={"ID":"9b39cef3-7078-4992-bd8c-7bbb57193b12","Type":"ContainerDied","Data":"d51cc11bfb8cfe13b889ea42b0148614c4afe76580d5a2d8107e7de382089bc8"} Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.938589 4894 scope.go:117] "RemoveContainer" containerID="df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.938592 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gnr8s" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.958749 4894 scope.go:117] "RemoveContainer" containerID="cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.985143 4894 scope.go:117] "RemoveContainer" containerID="8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31" Dec 08 15:58:36 crc kubenswrapper[4894]: I1208 15:58:36.995947 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnr8s"] Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.005109 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gnr8s"] Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.025892 4894 scope.go:117] "RemoveContainer" containerID="df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890" Dec 08 15:58:37 crc kubenswrapper[4894]: E1208 15:58:37.026381 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890\": container with ID starting with df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890 not found: ID does not exist" containerID="df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890" Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.026424 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890"} err="failed to get container status \"df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890\": rpc error: code = NotFound desc = could not find container \"df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890\": container with ID starting with df01100111bd9cb408c4c08e95a24b3bb248d41db7c48a7cfc7e6013f24f8890 not found: ID does not exist" Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.026456 4894 scope.go:117] "RemoveContainer" containerID="cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23" Dec 08 15:58:37 crc kubenswrapper[4894]: E1208 15:58:37.026905 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23\": container with ID starting with cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23 not found: ID does not exist" containerID="cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23" Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.026994 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23"} err="failed to get container status \"cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23\": rpc error: code = NotFound desc = could not find container \"cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23\": container with ID starting with cad3abb95611c8b8e223fe629a094361a2c573853096e34c2c5cf6cc96fc9c23 not found: ID does not exist" Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.027080 4894 scope.go:117] "RemoveContainer" containerID="8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31" Dec 08 15:58:37 crc kubenswrapper[4894]: E1208 15:58:37.027408 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31\": container with ID starting with 8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31 not found: ID does not exist" containerID="8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31" Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.027490 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31"} err="failed to get container status \"8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31\": rpc error: code = NotFound desc = could not find container \"8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31\": container with ID starting with 8318368ea4a84d59c690bfd2527d645d65536b29773c7e5ee248c694c2ca3b31 not found: ID does not exist" Dec 08 15:58:37 crc kubenswrapper[4894]: I1208 15:58:37.211634 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" path="/var/lib/kubelet/pods/9b39cef3-7078-4992-bd8c-7bbb57193b12/volumes" Dec 08 15:58:41 crc kubenswrapper[4894]: I1208 15:58:41.990179 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wtjnm"] Dec 08 15:58:41 crc kubenswrapper[4894]: E1208 15:58:41.996140 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="extract-utilities" Dec 08 15:58:41 crc kubenswrapper[4894]: I1208 15:58:41.996172 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="extract-utilities" Dec 08 15:58:41 crc kubenswrapper[4894]: E1208 15:58:41.996210 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="registry-server" Dec 08 15:58:41 crc kubenswrapper[4894]: I1208 15:58:41.996220 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="registry-server" Dec 08 15:58:41 crc kubenswrapper[4894]: E1208 15:58:41.996237 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="extract-content" Dec 08 15:58:41 crc kubenswrapper[4894]: I1208 15:58:41.996244 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="extract-content" Dec 08 15:58:41 crc kubenswrapper[4894]: I1208 15:58:41.996537 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b39cef3-7078-4992-bd8c-7bbb57193b12" containerName="registry-server" Dec 08 15:58:41 crc kubenswrapper[4894]: I1208 15:58:41.998329 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.003752 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtjnm"] Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.151280 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm5tv\" (UniqueName: \"kubernetes.io/projected/1b631db8-8ffe-4573-b002-b9162d4ca073-kube-api-access-nm5tv\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.151397 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b631db8-8ffe-4573-b002-b9162d4ca073-utilities\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.151471 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b631db8-8ffe-4573-b002-b9162d4ca073-catalog-content\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.253666 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm5tv\" (UniqueName: \"kubernetes.io/projected/1b631db8-8ffe-4573-b002-b9162d4ca073-kube-api-access-nm5tv\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.253891 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b631db8-8ffe-4573-b002-b9162d4ca073-utilities\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.254088 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b631db8-8ffe-4573-b002-b9162d4ca073-catalog-content\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.254911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b631db8-8ffe-4573-b002-b9162d4ca073-catalog-content\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.255269 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b631db8-8ffe-4573-b002-b9162d4ca073-utilities\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.524119 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm5tv\" (UniqueName: \"kubernetes.io/projected/1b631db8-8ffe-4573-b002-b9162d4ca073-kube-api-access-nm5tv\") pod \"redhat-operators-wtjnm\" (UID: \"1b631db8-8ffe-4573-b002-b9162d4ca073\") " pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:42 crc kubenswrapper[4894]: I1208 15:58:42.622502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjnm" Dec 08 15:58:43 crc kubenswrapper[4894]: I1208 15:58:43.177584 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtjnm"] Dec 08 15:58:43 crc kubenswrapper[4894]: I1208 15:58:43.998782 4894 generic.go:334] "Generic (PLEG): container finished" podID="1b631db8-8ffe-4573-b002-b9162d4ca073" containerID="979629d7a21744ce7dca4864a54e9b7907044ea69ed63ac3a5098d166e4041f6" exitCode=0 Dec 08 15:58:43 crc kubenswrapper[4894]: I1208 15:58:43.998841 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjnm" event={"ID":"1b631db8-8ffe-4573-b002-b9162d4ca073","Type":"ContainerDied","Data":"979629d7a21744ce7dca4864a54e9b7907044ea69ed63ac3a5098d166e4041f6"} Dec 08 15:58:43 crc kubenswrapper[4894]: I1208 15:58:43.999176 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjnm" event={"ID":"1b631db8-8ffe-4573-b002-b9162d4ca073","Type":"ContainerStarted","Data":"f7155bcf6349d28e34bc0836f07f401d20905b35b6a62c0a71dc7a8ae5970e70"} Dec 08 15:58:45 crc kubenswrapper[4894]: I1208 15:58:45.196606 4894 scope.go:117] "RemoveContainer" containerID="e99522aca5f18762d2e56ef5531f0d7acaa067e6bd6137a1c74afdea7f966df5" Dec 08 15:58:45 crc kubenswrapper[4894]: E1208 15:58:45.197072 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-97dqr_openshift-machine-config-operator(b27019e5-2a3d-414e-b2ee-7606492ba074)\"" pod="openshift-machine-config-operator/machine-config-daemon-97dqr" podUID="b27019e5-2a3d-414e-b2ee-7606492ba074" Dec 08 15:58:52 crc kubenswrapper[4894]: I1208 15:58:52.072107 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjnm" event={"ID":"1b631db8-8ffe-4573-b002-b9162d4ca073","Type":"ContainerStarted","Data":"3f8e0d3f849e5e9a1e6cd06a02ac80b998fd7bfd8c1ce0e64ed10c2fb9cf3c99"} Dec 08 15:58:53 crc kubenswrapper[4894]: I1208 15:58:53.084068 4894 generic.go:334] "Generic (PLEG): container finished" podID="1b631db8-8ffe-4573-b002-b9162d4ca073" containerID="3f8e0d3f849e5e9a1e6cd06a02ac80b998fd7bfd8c1ce0e64ed10c2fb9cf3c99" exitCode=0 Dec 08 15:58:53 crc kubenswrapper[4894]: I1208 15:58:53.084189 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjnm" event={"ID":"1b631db8-8ffe-4573-b002-b9162d4ca073","Type":"ContainerDied","Data":"3f8e0d3f849e5e9a1e6cd06a02ac80b998fd7bfd8c1ce0e64ed10c2fb9cf3c99"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515115572704024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015115572704017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015115561731016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015115561731015462 5ustar corecore